| /* |
| SPARC translation |
| |
| Copyright (C) 2003 Thomas M. Ogrisegg <tom@fnord.at> |
| Copyright (C) 2003-2005 Fabrice Bellard |
| |
| This library is free software; you can redistribute it and/or |
| modify it under the terms of the GNU Lesser General Public |
| License as published by the Free Software Foundation; either |
| version 2.1 of the License, or (at your option) any later version. |
| |
| This library is distributed in the hope that it will be useful, |
| but WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| Lesser General Public License for more details. |
| |
| You should have received a copy of the GNU Lesser General Public |
| License along with this library; if not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #include "qemu/osdep.h" |
| |
| #include "cpu.h" |
| #include "exec/helper-proto.h" |
| #include "exec/exec-all.h" |
| #include "tcg/tcg-op.h" |
| #include "tcg/tcg-op-gvec.h" |
| #include "exec/helper-gen.h" |
| #include "exec/translator.h" |
| #include "exec/log.h" |
| #include "fpu/softfloat.h" |
| #include "asi.h" |
| |
| #define HELPER_H "helper.h" |
| #include "exec/helper-info.c.inc" |
| #undef HELPER_H |
| |
| #ifdef TARGET_SPARC64 |
| # define gen_helper_rdpsr(D, E) qemu_build_not_reached() |
| # define gen_helper_rdasr17(D, E) qemu_build_not_reached() |
| # define gen_helper_rett(E) qemu_build_not_reached() |
| # define gen_helper_power_down(E) qemu_build_not_reached() |
| # define gen_helper_wrpsr(E, S) qemu_build_not_reached() |
| #else |
| # define gen_helper_clear_softint(E, S) qemu_build_not_reached() |
| # define gen_helper_done(E) qemu_build_not_reached() |
| # define gen_helper_flushw(E) qemu_build_not_reached() |
| # define gen_helper_fmul8x16a(D, S1, S2) qemu_build_not_reached() |
| # define gen_helper_rdccr(D, E) qemu_build_not_reached() |
| # define gen_helper_rdcwp(D, E) qemu_build_not_reached() |
| # define gen_helper_restored(E) qemu_build_not_reached() |
| # define gen_helper_retry(E) qemu_build_not_reached() |
| # define gen_helper_saved(E) qemu_build_not_reached() |
| # define gen_helper_set_softint(E, S) qemu_build_not_reached() |
| # define gen_helper_tick_get_count(D, E, T, C) qemu_build_not_reached() |
| # define gen_helper_tick_set_count(P, S) qemu_build_not_reached() |
| # define gen_helper_tick_set_limit(P, S) qemu_build_not_reached() |
| # define gen_helper_wrccr(E, S) qemu_build_not_reached() |
| # define gen_helper_wrcwp(E, S) qemu_build_not_reached() |
| # define gen_helper_wrgl(E, S) qemu_build_not_reached() |
| # define gen_helper_write_softint(E, S) qemu_build_not_reached() |
| # define gen_helper_wrpil(E, S) qemu_build_not_reached() |
| # define gen_helper_wrpstate(E, S) qemu_build_not_reached() |
| # define gen_helper_cmask8 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_cmask16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_cmask32 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpeq8 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpeq16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpeq32 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpgt8 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpgt16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpgt32 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmple8 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmple16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmple32 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpne8 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpne16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpne32 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpule8 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpule16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpule32 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpugt8 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpugt16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fcmpugt32 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fdtox ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fexpand ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fmul8sux16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fmul8ulx16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fmul8x16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fpmerge ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fqtox ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fslas16 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fslas32 ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fstox ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fxtod ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fxtoq ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_fxtos ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_pdist ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_xmulx ({ qemu_build_not_reached(); NULL; }) |
| # define gen_helper_xmulxhi ({ qemu_build_not_reached(); NULL; }) |
| # define MAXTL_MASK 0 |
| #endif |
| |
| /* Dynamic PC, must exit to main loop. */ |
| #define DYNAMIC_PC 1 |
| /* Dynamic PC, one of two values according to jump_pc[T2]. */ |
| #define JUMP_PC 2 |
| /* Dynamic PC, may lookup next TB. */ |
| #define DYNAMIC_PC_LOOKUP 3 |
| |
| #define DISAS_EXIT DISAS_TARGET_0 |
| |
| /* global register indexes */ |
| static TCGv_ptr cpu_regwptr; |
| static TCGv cpu_pc, cpu_npc; |
| static TCGv cpu_regs[32]; |
| static TCGv cpu_y; |
| static TCGv cpu_tbr; |
| static TCGv cpu_cond; |
| static TCGv cpu_cc_N; |
| static TCGv cpu_cc_V; |
| static TCGv cpu_icc_Z; |
| static TCGv cpu_icc_C; |
| #ifdef TARGET_SPARC64 |
| static TCGv cpu_xcc_Z; |
| static TCGv cpu_xcc_C; |
| static TCGv_i32 cpu_fprs; |
| static TCGv cpu_gsr; |
| #else |
| # define cpu_fprs ({ qemu_build_not_reached(); (TCGv)NULL; }) |
| # define cpu_gsr ({ qemu_build_not_reached(); (TCGv)NULL; }) |
| #endif |
| |
| #ifdef TARGET_SPARC64 |
| #define cpu_cc_Z cpu_xcc_Z |
| #define cpu_cc_C cpu_xcc_C |
| #else |
| #define cpu_cc_Z cpu_icc_Z |
| #define cpu_cc_C cpu_icc_C |
| #define cpu_xcc_Z ({ qemu_build_not_reached(); NULL; }) |
| #define cpu_xcc_C ({ qemu_build_not_reached(); NULL; }) |
| #endif |
| |
| /* Floating point comparison registers */ |
| static TCGv_i32 cpu_fcc[TARGET_FCCREGS]; |
| |
| #define env_field_offsetof(X) offsetof(CPUSPARCState, X) |
| #ifdef TARGET_SPARC64 |
| # define env32_field_offsetof(X) ({ qemu_build_not_reached(); 0; }) |
| # define env64_field_offsetof(X) env_field_offsetof(X) |
| #else |
| # define env32_field_offsetof(X) env_field_offsetof(X) |
| # define env64_field_offsetof(X) ({ qemu_build_not_reached(); 0; }) |
| #endif |
| |
| typedef struct DisasCompare { |
| TCGCond cond; |
| TCGv c1; |
| int c2; |
| } DisasCompare; |
| |
| typedef struct DisasDelayException { |
| struct DisasDelayException *next; |
| TCGLabel *lab; |
| TCGv_i32 excp; |
| /* Saved state at parent insn. */ |
| target_ulong pc; |
| target_ulong npc; |
| } DisasDelayException; |
| |
| typedef struct DisasContext { |
| DisasContextBase base; |
| target_ulong pc; /* current Program Counter: integer or DYNAMIC_PC */ |
| target_ulong npc; /* next PC: integer or DYNAMIC_PC or JUMP_PC */ |
| |
| /* Used when JUMP_PC value is used. */ |
| DisasCompare jump; |
| target_ulong jump_pc[2]; |
| |
| int mem_idx; |
| bool cpu_cond_live; |
| bool fpu_enabled; |
| bool address_mask_32bit; |
| #ifndef CONFIG_USER_ONLY |
| bool supervisor; |
| #ifdef TARGET_SPARC64 |
| bool hypervisor; |
| #endif |
| #endif |
| |
| sparc_def_t *def; |
| #ifdef TARGET_SPARC64 |
| int fprs_dirty; |
| int asi; |
| #endif |
| DisasDelayException *delay_excp_list; |
| } DisasContext; |
| |
| // This function uses non-native bit order |
| #define GET_FIELD(X, FROM, TO) \ |
| ((X) >> (31 - (TO)) & ((1 << ((TO) - (FROM) + 1)) - 1)) |
| |
| // This function uses the order in the manuals, i.e. bit 0 is 2^0 |
| #define GET_FIELD_SP(X, FROM, TO) \ |
| GET_FIELD(X, 31 - (TO), 31 - (FROM)) |
| |
| #define GET_FIELDs(x,a,b) sign_extend (GET_FIELD(x,a,b), (b) - (a) + 1) |
| #define GET_FIELD_SPs(x,a,b) sign_extend (GET_FIELD_SP(x,a,b), ((b) - (a) + 1)) |
| |
| #define UA2005_HTRAP_MASK 0xff |
| #define V8_TRAP_MASK 0x7f |
| |
| #define IS_IMM (insn & (1<<13)) |
| |
| static void gen_update_fprs_dirty(DisasContext *dc, int rd) |
| { |
| #if defined(TARGET_SPARC64) |
| int bit = (rd < 32) ? 1 : 2; |
| /* If we know we've already set this bit within the TB, |
| we can avoid setting it again. */ |
| if (!(dc->fprs_dirty & bit)) { |
| dc->fprs_dirty |= bit; |
| tcg_gen_ori_i32(cpu_fprs, cpu_fprs, bit); |
| } |
| #endif |
| } |
| |
| /* floating point registers moves */ |
| |
| static int gen_offset_fpr_F(unsigned int reg) |
| { |
| int ret; |
| |
| tcg_debug_assert(reg < 32); |
| ret= offsetof(CPUSPARCState, fpr[reg / 2]); |
| if (reg & 1) { |
| ret += offsetof(CPU_DoubleU, l.lower); |
| } else { |
| ret += offsetof(CPU_DoubleU, l.upper); |
| } |
| return ret; |
| } |
| |
| static TCGv_i32 gen_load_fpr_F(DisasContext *dc, unsigned int src) |
| { |
| TCGv_i32 ret = tcg_temp_new_i32(); |
| tcg_gen_ld_i32(ret, tcg_env, gen_offset_fpr_F(src)); |
| return ret; |
| } |
| |
| static void gen_store_fpr_F(DisasContext *dc, unsigned int dst, TCGv_i32 v) |
| { |
| tcg_gen_st_i32(v, tcg_env, gen_offset_fpr_F(dst)); |
| gen_update_fprs_dirty(dc, dst); |
| } |
| |
| static int gen_offset_fpr_D(unsigned int reg) |
| { |
| tcg_debug_assert(reg < 64); |
| tcg_debug_assert(reg % 2 == 0); |
| return offsetof(CPUSPARCState, fpr[reg / 2]); |
| } |
| |
| static TCGv_i64 gen_load_fpr_D(DisasContext *dc, unsigned int src) |
| { |
| TCGv_i64 ret = tcg_temp_new_i64(); |
| tcg_gen_ld_i64(ret, tcg_env, gen_offset_fpr_D(src)); |
| return ret; |
| } |
| |
| static void gen_store_fpr_D(DisasContext *dc, unsigned int dst, TCGv_i64 v) |
| { |
| tcg_gen_st_i64(v, tcg_env, gen_offset_fpr_D(dst)); |
| gen_update_fprs_dirty(dc, dst); |
| } |
| |
| static TCGv_i128 gen_load_fpr_Q(DisasContext *dc, unsigned int src) |
| { |
| TCGv_i128 ret = tcg_temp_new_i128(); |
| TCGv_i64 h = gen_load_fpr_D(dc, src); |
| TCGv_i64 l = gen_load_fpr_D(dc, src + 2); |
| |
| tcg_gen_concat_i64_i128(ret, l, h); |
| return ret; |
| } |
| |
| static void gen_store_fpr_Q(DisasContext *dc, unsigned int dst, TCGv_i128 v) |
| { |
| TCGv_i64 h = tcg_temp_new_i64(); |
| TCGv_i64 l = tcg_temp_new_i64(); |
| |
| tcg_gen_extr_i128_i64(l, h, v); |
| gen_store_fpr_D(dc, dst, h); |
| gen_store_fpr_D(dc, dst + 2, l); |
| } |
| |
| /* moves */ |
| #ifdef CONFIG_USER_ONLY |
| #define supervisor(dc) 0 |
| #define hypervisor(dc) 0 |
| #else |
| #ifdef TARGET_SPARC64 |
| #define hypervisor(dc) (dc->hypervisor) |
| #define supervisor(dc) (dc->supervisor | dc->hypervisor) |
| #else |
| #define supervisor(dc) (dc->supervisor) |
| #define hypervisor(dc) 0 |
| #endif |
| #endif |
| |
| #if !defined(TARGET_SPARC64) |
| # define AM_CHECK(dc) false |
| #elif defined(TARGET_ABI32) |
| # define AM_CHECK(dc) true |
| #elif defined(CONFIG_USER_ONLY) |
| # define AM_CHECK(dc) false |
| #else |
| # define AM_CHECK(dc) ((dc)->address_mask_32bit) |
| #endif |
| |
| static void gen_address_mask(DisasContext *dc, TCGv addr) |
| { |
| if (AM_CHECK(dc)) { |
| tcg_gen_andi_tl(addr, addr, 0xffffffffULL); |
| } |
| } |
| |
| static target_ulong address_mask_i(DisasContext *dc, target_ulong addr) |
| { |
| return AM_CHECK(dc) ? (uint32_t)addr : addr; |
| } |
| |
| static TCGv gen_load_gpr(DisasContext *dc, int reg) |
| { |
| if (reg > 0) { |
| assert(reg < 32); |
| return cpu_regs[reg]; |
| } else { |
| TCGv t = tcg_temp_new(); |
| tcg_gen_movi_tl(t, 0); |
| return t; |
| } |
| } |
| |
| static void gen_store_gpr(DisasContext *dc, int reg, TCGv v) |
| { |
| if (reg > 0) { |
| assert(reg < 32); |
| tcg_gen_mov_tl(cpu_regs[reg], v); |
| } |
| } |
| |
| static TCGv gen_dest_gpr(DisasContext *dc, int reg) |
| { |
| if (reg > 0) { |
| assert(reg < 32); |
| return cpu_regs[reg]; |
| } else { |
| return tcg_temp_new(); |
| } |
| } |
| |
| static bool use_goto_tb(DisasContext *s, target_ulong pc, target_ulong npc) |
| { |
| return translator_use_goto_tb(&s->base, pc) && |
| translator_use_goto_tb(&s->base, npc); |
| } |
| |
| static void gen_goto_tb(DisasContext *s, int tb_num, |
| target_ulong pc, target_ulong npc) |
| { |
| if (use_goto_tb(s, pc, npc)) { |
| /* jump to same page: we can use a direct jump */ |
| tcg_gen_goto_tb(tb_num); |
| tcg_gen_movi_tl(cpu_pc, pc); |
| tcg_gen_movi_tl(cpu_npc, npc); |
| tcg_gen_exit_tb(s->base.tb, tb_num); |
| } else { |
| /* jump to another page: we can use an indirect jump */ |
| tcg_gen_movi_tl(cpu_pc, pc); |
| tcg_gen_movi_tl(cpu_npc, npc); |
| tcg_gen_lookup_and_goto_ptr(); |
| } |
| } |
| |
| static TCGv gen_carry32(void) |
| { |
| if (TARGET_LONG_BITS == 64) { |
| TCGv t = tcg_temp_new(); |
| tcg_gen_extract_tl(t, cpu_icc_C, 32, 1); |
| return t; |
| } |
| return cpu_icc_C; |
| } |
| |
| static void gen_op_addcc_int(TCGv dst, TCGv src1, TCGv src2, TCGv cin) |
| { |
| TCGv z = tcg_constant_tl(0); |
| |
| if (cin) { |
| tcg_gen_add2_tl(cpu_cc_N, cpu_cc_C, src1, z, cin, z); |
| tcg_gen_add2_tl(cpu_cc_N, cpu_cc_C, cpu_cc_N, cpu_cc_C, src2, z); |
| } else { |
| tcg_gen_add2_tl(cpu_cc_N, cpu_cc_C, src1, z, src2, z); |
| } |
| tcg_gen_xor_tl(cpu_cc_Z, src1, src2); |
| tcg_gen_xor_tl(cpu_cc_V, cpu_cc_N, src2); |
| tcg_gen_andc_tl(cpu_cc_V, cpu_cc_V, cpu_cc_Z); |
| if (TARGET_LONG_BITS == 64) { |
| /* |
| * Carry-in to bit 32 is result ^ src1 ^ src2. |
| * We already have the src xor term in Z, from computation of V. |
| */ |
| tcg_gen_xor_tl(cpu_icc_C, cpu_cc_Z, cpu_cc_N); |
| tcg_gen_mov_tl(cpu_icc_Z, cpu_cc_N); |
| } |
| tcg_gen_mov_tl(cpu_cc_Z, cpu_cc_N); |
| tcg_gen_mov_tl(dst, cpu_cc_N); |
| } |
| |
| static void gen_op_addcc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_op_addcc_int(dst, src1, src2, NULL); |
| } |
| |
| static void gen_op_taddcc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| TCGv t = tcg_temp_new(); |
| |
| /* Save the tag bits around modification of dst. */ |
| tcg_gen_or_tl(t, src1, src2); |
| |
| gen_op_addcc(dst, src1, src2); |
| |
| /* Incorprate tag bits into icc.V */ |
| tcg_gen_andi_tl(t, t, 3); |
| tcg_gen_neg_tl(t, t); |
| tcg_gen_ext32u_tl(t, t); |
| tcg_gen_or_tl(cpu_cc_V, cpu_cc_V, t); |
| } |
| |
| static void gen_op_addc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| tcg_gen_add_tl(dst, src1, src2); |
| tcg_gen_add_tl(dst, dst, gen_carry32()); |
| } |
| |
| static void gen_op_addccc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_op_addcc_int(dst, src1, src2, gen_carry32()); |
| } |
| |
| static void gen_op_addxc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| tcg_gen_add_tl(dst, src1, src2); |
| tcg_gen_add_tl(dst, dst, cpu_cc_C); |
| } |
| |
| static void gen_op_addxccc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_op_addcc_int(dst, src1, src2, cpu_cc_C); |
| } |
| |
| static void gen_op_subcc_int(TCGv dst, TCGv src1, TCGv src2, TCGv cin) |
| { |
| TCGv z = tcg_constant_tl(0); |
| |
| if (cin) { |
| tcg_gen_sub2_tl(cpu_cc_N, cpu_cc_C, src1, z, cin, z); |
| tcg_gen_sub2_tl(cpu_cc_N, cpu_cc_C, cpu_cc_N, cpu_cc_C, src2, z); |
| } else { |
| tcg_gen_sub2_tl(cpu_cc_N, cpu_cc_C, src1, z, src2, z); |
| } |
| tcg_gen_neg_tl(cpu_cc_C, cpu_cc_C); |
| tcg_gen_xor_tl(cpu_cc_Z, src1, src2); |
| tcg_gen_xor_tl(cpu_cc_V, cpu_cc_N, src1); |
| tcg_gen_and_tl(cpu_cc_V, cpu_cc_V, cpu_cc_Z); |
| #ifdef TARGET_SPARC64 |
| tcg_gen_xor_tl(cpu_icc_C, cpu_cc_Z, cpu_cc_N); |
| tcg_gen_mov_tl(cpu_icc_Z, cpu_cc_N); |
| #endif |
| tcg_gen_mov_tl(cpu_cc_Z, cpu_cc_N); |
| tcg_gen_mov_tl(dst, cpu_cc_N); |
| } |
| |
| static void gen_op_subcc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_op_subcc_int(dst, src1, src2, NULL); |
| } |
| |
| static void gen_op_tsubcc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| TCGv t = tcg_temp_new(); |
| |
| /* Save the tag bits around modification of dst. */ |
| tcg_gen_or_tl(t, src1, src2); |
| |
| gen_op_subcc(dst, src1, src2); |
| |
| /* Incorprate tag bits into icc.V */ |
| tcg_gen_andi_tl(t, t, 3); |
| tcg_gen_neg_tl(t, t); |
| tcg_gen_ext32u_tl(t, t); |
| tcg_gen_or_tl(cpu_cc_V, cpu_cc_V, t); |
| } |
| |
| static void gen_op_subc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| tcg_gen_sub_tl(dst, src1, src2); |
| tcg_gen_sub_tl(dst, dst, gen_carry32()); |
| } |
| |
| static void gen_op_subccc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_op_subcc_int(dst, src1, src2, gen_carry32()); |
| } |
| |
| static void gen_op_subxc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| tcg_gen_sub_tl(dst, src1, src2); |
| tcg_gen_sub_tl(dst, dst, cpu_cc_C); |
| } |
| |
| static void gen_op_subxccc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_op_subcc_int(dst, src1, src2, cpu_cc_C); |
| } |
| |
| static void gen_op_mulscc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| TCGv zero = tcg_constant_tl(0); |
| TCGv one = tcg_constant_tl(1); |
| TCGv t_src1 = tcg_temp_new(); |
| TCGv t_src2 = tcg_temp_new(); |
| TCGv t0 = tcg_temp_new(); |
| |
| tcg_gen_ext32u_tl(t_src1, src1); |
| tcg_gen_ext32u_tl(t_src2, src2); |
| |
| /* |
| * if (!(env->y & 1)) |
| * src2 = 0; |
| */ |
| tcg_gen_movcond_tl(TCG_COND_TSTEQ, t_src2, cpu_y, one, zero, t_src2); |
| |
| /* |
| * b2 = src1 & 1; |
| * y = (b2 << 31) | (y >> 1); |
| */ |
| tcg_gen_extract_tl(t0, cpu_y, 1, 31); |
| tcg_gen_deposit_tl(cpu_y, t0, src1, 31, 1); |
| |
| // b1 = N ^ V; |
| tcg_gen_xor_tl(t0, cpu_cc_N, cpu_cc_V); |
| |
| /* |
| * src1 = (b1 << 31) | (src1 >> 1) |
| */ |
| tcg_gen_andi_tl(t0, t0, 1u << 31); |
| tcg_gen_shri_tl(t_src1, t_src1, 1); |
| tcg_gen_or_tl(t_src1, t_src1, t0); |
| |
| gen_op_addcc(dst, t_src1, t_src2); |
| } |
| |
| static void gen_op_multiply(TCGv dst, TCGv src1, TCGv src2, int sign_ext) |
| { |
| #if TARGET_LONG_BITS == 32 |
| if (sign_ext) { |
| tcg_gen_muls2_tl(dst, cpu_y, src1, src2); |
| } else { |
| tcg_gen_mulu2_tl(dst, cpu_y, src1, src2); |
| } |
| #else |
| TCGv t0 = tcg_temp_new_i64(); |
| TCGv t1 = tcg_temp_new_i64(); |
| |
| if (sign_ext) { |
| tcg_gen_ext32s_i64(t0, src1); |
| tcg_gen_ext32s_i64(t1, src2); |
| } else { |
| tcg_gen_ext32u_i64(t0, src1); |
| tcg_gen_ext32u_i64(t1, src2); |
| } |
| |
| tcg_gen_mul_i64(dst, t0, t1); |
| tcg_gen_shri_i64(cpu_y, dst, 32); |
| #endif |
| } |
| |
| static void gen_op_umul(TCGv dst, TCGv src1, TCGv src2) |
| { |
| /* zero-extend truncated operands before multiplication */ |
| gen_op_multiply(dst, src1, src2, 0); |
| } |
| |
| static void gen_op_smul(TCGv dst, TCGv src1, TCGv src2) |
| { |
| /* sign-extend truncated operands before multiplication */ |
| gen_op_multiply(dst, src1, src2, 1); |
| } |
| |
| static void gen_op_umulxhi(TCGv dst, TCGv src1, TCGv src2) |
| { |
| TCGv discard = tcg_temp_new(); |
| tcg_gen_mulu2_tl(discard, dst, src1, src2); |
| } |
| |
| static void gen_op_fpmaddx(TCGv_i64 dst, TCGv_i64 src1, |
| TCGv_i64 src2, TCGv_i64 src3) |
| { |
| TCGv_i64 t = tcg_temp_new_i64(); |
| |
| tcg_gen_mul_i64(t, src1, src2); |
| tcg_gen_add_i64(dst, src3, t); |
| } |
| |
| static void gen_op_fpmaddxhi(TCGv_i64 dst, TCGv_i64 src1, |
| TCGv_i64 src2, TCGv_i64 src3) |
| { |
| TCGv_i64 l = tcg_temp_new_i64(); |
| TCGv_i64 h = tcg_temp_new_i64(); |
| TCGv_i64 z = tcg_constant_i64(0); |
| |
| tcg_gen_mulu2_i64(l, h, src1, src2); |
| tcg_gen_add2_i64(l, dst, l, h, src3, z); |
| } |
| |
| static void gen_op_sdiv(TCGv dst, TCGv src1, TCGv src2) |
| { |
| #ifdef TARGET_SPARC64 |
| gen_helper_sdiv(dst, tcg_env, src1, src2); |
| tcg_gen_ext32s_tl(dst, dst); |
| #else |
| TCGv_i64 t64 = tcg_temp_new_i64(); |
| gen_helper_sdiv(t64, tcg_env, src1, src2); |
| tcg_gen_trunc_i64_tl(dst, t64); |
| #endif |
| } |
| |
| static void gen_op_udivcc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| TCGv_i64 t64; |
| |
| #ifdef TARGET_SPARC64 |
| t64 = cpu_cc_V; |
| #else |
| t64 = tcg_temp_new_i64(); |
| #endif |
| |
| gen_helper_udiv(t64, tcg_env, src1, src2); |
| |
| #ifdef TARGET_SPARC64 |
| tcg_gen_ext32u_tl(cpu_cc_N, t64); |
| tcg_gen_shri_tl(cpu_cc_V, t64, 32); |
| tcg_gen_mov_tl(cpu_icc_Z, cpu_cc_N); |
| tcg_gen_movi_tl(cpu_icc_C, 0); |
| #else |
| tcg_gen_extr_i64_tl(cpu_cc_N, cpu_cc_V, t64); |
| #endif |
| tcg_gen_mov_tl(cpu_cc_Z, cpu_cc_N); |
| tcg_gen_movi_tl(cpu_cc_C, 0); |
| tcg_gen_mov_tl(dst, cpu_cc_N); |
| } |
| |
| static void gen_op_sdivcc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| TCGv_i64 t64; |
| |
| #ifdef TARGET_SPARC64 |
| t64 = cpu_cc_V; |
| #else |
| t64 = tcg_temp_new_i64(); |
| #endif |
| |
| gen_helper_sdiv(t64, tcg_env, src1, src2); |
| |
| #ifdef TARGET_SPARC64 |
| tcg_gen_ext32s_tl(cpu_cc_N, t64); |
| tcg_gen_shri_tl(cpu_cc_V, t64, 32); |
| tcg_gen_mov_tl(cpu_icc_Z, cpu_cc_N); |
| tcg_gen_movi_tl(cpu_icc_C, 0); |
| #else |
| tcg_gen_extr_i64_tl(cpu_cc_N, cpu_cc_V, t64); |
| #endif |
| tcg_gen_mov_tl(cpu_cc_Z, cpu_cc_N); |
| tcg_gen_movi_tl(cpu_cc_C, 0); |
| tcg_gen_mov_tl(dst, cpu_cc_N); |
| } |
| |
| static void gen_op_taddcctv(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_helper_taddcctv(dst, tcg_env, src1, src2); |
| } |
| |
| static void gen_op_tsubcctv(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_helper_tsubcctv(dst, tcg_env, src1, src2); |
| } |
| |
| static void gen_op_popc(TCGv dst, TCGv src1, TCGv src2) |
| { |
| tcg_gen_ctpop_tl(dst, src2); |
| } |
| |
| static void gen_op_lzcnt(TCGv dst, TCGv src) |
| { |
| tcg_gen_clzi_tl(dst, src, TARGET_LONG_BITS); |
| } |
| |
| #ifndef TARGET_SPARC64 |
| static void gen_helper_array8(TCGv dst, TCGv src1, TCGv src2) |
| { |
| g_assert_not_reached(); |
| } |
| #endif |
| |
| static void gen_op_array16(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_helper_array8(dst, src1, src2); |
| tcg_gen_shli_tl(dst, dst, 1); |
| } |
| |
| static void gen_op_array32(TCGv dst, TCGv src1, TCGv src2) |
| { |
| gen_helper_array8(dst, src1, src2); |
| tcg_gen_shli_tl(dst, dst, 2); |
| } |
| |
| static void gen_op_fpack16(TCGv_i32 dst, TCGv_i64 src) |
| { |
| #ifdef TARGET_SPARC64 |
| gen_helper_fpack16(dst, cpu_gsr, src); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| static void gen_op_fpackfix(TCGv_i32 dst, TCGv_i64 src) |
| { |
| #ifdef TARGET_SPARC64 |
| gen_helper_fpackfix(dst, cpu_gsr, src); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| static void gen_op_fpack32(TCGv_i64 dst, TCGv_i64 src1, TCGv_i64 src2) |
| { |
| #ifdef TARGET_SPARC64 |
| gen_helper_fpack32(dst, cpu_gsr, src1, src2); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| static void gen_op_fpadds16s(TCGv_i32 d, TCGv_i32 src1, TCGv_i32 src2) |
| { |
| TCGv_i32 t[2]; |
| |
| for (int i = 0; i < 2; i++) { |
| TCGv_i32 u = tcg_temp_new_i32(); |
| TCGv_i32 v = tcg_temp_new_i32(); |
| |
| tcg_gen_sextract_i32(u, src1, i * 16, 16); |
| tcg_gen_sextract_i32(v, src2, i * 16, 16); |
| tcg_gen_add_i32(u, u, v); |
| tcg_gen_smax_i32(u, u, tcg_constant_i32(INT16_MIN)); |
| tcg_gen_smin_i32(u, u, tcg_constant_i32(INT16_MAX)); |
| t[i] = u; |
| } |
| tcg_gen_deposit_i32(d, t[0], t[1], 16, 16); |
| } |
| |
| static void gen_op_fpsubs16s(TCGv_i32 d, TCGv_i32 src1, TCGv_i32 src2) |
| { |
| TCGv_i32 t[2]; |
| |
| for (int i = 0; i < 2; i++) { |
| TCGv_i32 u = tcg_temp_new_i32(); |
| TCGv_i32 v = tcg_temp_new_i32(); |
| |
| tcg_gen_sextract_i32(u, src1, i * 16, 16); |
| tcg_gen_sextract_i32(v, src2, i * 16, 16); |
| tcg_gen_sub_i32(u, u, v); |
| tcg_gen_smax_i32(u, u, tcg_constant_i32(INT16_MIN)); |
| tcg_gen_smin_i32(u, u, tcg_constant_i32(INT16_MAX)); |
| t[i] = u; |
| } |
| tcg_gen_deposit_i32(d, t[0], t[1], 16, 16); |
| } |
| |
| static void gen_op_fpadds32s(TCGv_i32 d, TCGv_i32 src1, TCGv_i32 src2) |
| { |
| TCGv_i32 r = tcg_temp_new_i32(); |
| TCGv_i32 t = tcg_temp_new_i32(); |
| TCGv_i32 v = tcg_temp_new_i32(); |
| TCGv_i32 z = tcg_constant_i32(0); |
| |
| tcg_gen_add_i32(r, src1, src2); |
| tcg_gen_xor_i32(t, src1, src2); |
| tcg_gen_xor_i32(v, r, src2); |
| tcg_gen_andc_i32(v, v, t); |
| |
| tcg_gen_setcond_i32(TCG_COND_GE, t, r, z); |
| tcg_gen_addi_i32(t, t, INT32_MAX); |
| |
| tcg_gen_movcond_i32(TCG_COND_LT, d, v, z, t, r); |
| } |
| |
| static void gen_op_fpsubs32s(TCGv_i32 d, TCGv_i32 src1, TCGv_i32 src2) |
| { |
| TCGv_i32 r = tcg_temp_new_i32(); |
| TCGv_i32 t = tcg_temp_new_i32(); |
| TCGv_i32 v = tcg_temp_new_i32(); |
| TCGv_i32 z = tcg_constant_i32(0); |
| |
| tcg_gen_sub_i32(r, src1, src2); |
| tcg_gen_xor_i32(t, src1, src2); |
| tcg_gen_xor_i32(v, r, src1); |
| tcg_gen_and_i32(v, v, t); |
| |
| tcg_gen_setcond_i32(TCG_COND_GE, t, r, z); |
| tcg_gen_addi_i32(t, t, INT32_MAX); |
| |
| tcg_gen_movcond_i32(TCG_COND_LT, d, v, z, t, r); |
| } |
| |
| static void gen_op_faligndata_i(TCGv_i64 dst, TCGv_i64 s1, |
| TCGv_i64 s2, TCGv gsr) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv t1, t2, shift; |
| |
| t1 = tcg_temp_new(); |
| t2 = tcg_temp_new(); |
| shift = tcg_temp_new(); |
| |
| tcg_gen_andi_tl(shift, gsr, 7); |
| tcg_gen_shli_tl(shift, shift, 3); |
| tcg_gen_shl_tl(t1, s1, shift); |
| |
| /* |
| * A shift of 64 does not produce 0 in TCG. Divide this into a |
| * shift of (up to 63) followed by a constant shift of 1. |
| */ |
| tcg_gen_xori_tl(shift, shift, 63); |
| tcg_gen_shr_tl(t2, s2, shift); |
| tcg_gen_shri_tl(t2, t2, 1); |
| |
| tcg_gen_or_tl(dst, t1, t2); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| static void gen_op_faligndata_g(TCGv_i64 dst, TCGv_i64 s1, TCGv_i64 s2) |
| { |
| gen_op_faligndata_i(dst, s1, s2, cpu_gsr); |
| } |
| |
| static void gen_op_bshuffle(TCGv_i64 dst, TCGv_i64 src1, TCGv_i64 src2) |
| { |
| #ifdef TARGET_SPARC64 |
| gen_helper_bshuffle(dst, cpu_gsr, src1, src2); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| static void gen_op_pdistn(TCGv dst, TCGv_i64 src1, TCGv_i64 src2) |
| { |
| #ifdef TARGET_SPARC64 |
| gen_helper_pdist(dst, tcg_constant_i64(0), src1, src2); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| static void gen_op_fmul8x16al(TCGv_i64 dst, TCGv_i32 src1, TCGv_i32 src2) |
| { |
| tcg_gen_ext16s_i32(src2, src2); |
| gen_helper_fmul8x16a(dst, src1, src2); |
| } |
| |
| static void gen_op_fmul8x16au(TCGv_i64 dst, TCGv_i32 src1, TCGv_i32 src2) |
| { |
| tcg_gen_sari_i32(src2, src2, 16); |
| gen_helper_fmul8x16a(dst, src1, src2); |
| } |
| |
| static void gen_op_fmuld8ulx16(TCGv_i64 dst, TCGv_i32 src1, TCGv_i32 src2) |
| { |
| TCGv_i32 t0 = tcg_temp_new_i32(); |
| TCGv_i32 t1 = tcg_temp_new_i32(); |
| TCGv_i32 t2 = tcg_temp_new_i32(); |
| |
| tcg_gen_ext8u_i32(t0, src1); |
| tcg_gen_ext16s_i32(t1, src2); |
| tcg_gen_mul_i32(t0, t0, t1); |
| |
| tcg_gen_extract_i32(t1, src1, 16, 8); |
| tcg_gen_sextract_i32(t2, src2, 16, 16); |
| tcg_gen_mul_i32(t1, t1, t2); |
| |
| tcg_gen_concat_i32_i64(dst, t0, t1); |
| } |
| |
| static void gen_op_fmuld8sux16(TCGv_i64 dst, TCGv_i32 src1, TCGv_i32 src2) |
| { |
| TCGv_i32 t0 = tcg_temp_new_i32(); |
| TCGv_i32 t1 = tcg_temp_new_i32(); |
| TCGv_i32 t2 = tcg_temp_new_i32(); |
| |
| /* |
| * The insn description talks about extracting the upper 8 bits |
| * of the signed 16-bit input rs1, performing the multiply, then |
| * shifting left by 8 bits. Instead, zap the lower 8 bits of |
| * the rs1 input, which avoids the need for two shifts. |
| */ |
| tcg_gen_ext16s_i32(t0, src1); |
| tcg_gen_andi_i32(t0, t0, ~0xff); |
| tcg_gen_ext16s_i32(t1, src2); |
| tcg_gen_mul_i32(t0, t0, t1); |
| |
| tcg_gen_sextract_i32(t1, src1, 16, 16); |
| tcg_gen_andi_i32(t1, t1, ~0xff); |
| tcg_gen_sextract_i32(t2, src2, 16, 16); |
| tcg_gen_mul_i32(t1, t1, t2); |
| |
| tcg_gen_concat_i32_i64(dst, t0, t1); |
| } |
| |
| #ifdef TARGET_SPARC64 |
| static void gen_vec_fchksm16(unsigned vece, TCGv_vec dst, |
| TCGv_vec src1, TCGv_vec src2) |
| { |
| TCGv_vec a = tcg_temp_new_vec_matching(dst); |
| TCGv_vec c = tcg_temp_new_vec_matching(dst); |
| |
| tcg_gen_add_vec(vece, a, src1, src2); |
| tcg_gen_cmp_vec(TCG_COND_LTU, vece, c, a, src1); |
| /* Vector cmp produces -1 for true, so subtract to add carry. */ |
| tcg_gen_sub_vec(vece, dst, a, c); |
| } |
| |
| static void gen_op_fchksm16(unsigned vece, uint32_t dofs, uint32_t aofs, |
| uint32_t bofs, uint32_t oprsz, uint32_t maxsz) |
| { |
| static const TCGOpcode vecop_list[] = { |
| INDEX_op_cmp_vec, INDEX_op_add_vec, INDEX_op_sub_vec, |
| }; |
| static const GVecGen3 op = { |
| .fni8 = gen_helper_fchksm16, |
| .fniv = gen_vec_fchksm16, |
| .opt_opc = vecop_list, |
| .vece = MO_16, |
| }; |
| tcg_gen_gvec_3(dofs, aofs, bofs, oprsz, maxsz, &op); |
| } |
| |
| static void gen_vec_fmean16(unsigned vece, TCGv_vec dst, |
| TCGv_vec src1, TCGv_vec src2) |
| { |
| TCGv_vec t = tcg_temp_new_vec_matching(dst); |
| |
| tcg_gen_or_vec(vece, t, src1, src2); |
| tcg_gen_and_vec(vece, t, t, tcg_constant_vec_matching(dst, vece, 1)); |
| tcg_gen_sari_vec(vece, src1, src1, 1); |
| tcg_gen_sari_vec(vece, src2, src2, 1); |
| tcg_gen_add_vec(vece, dst, src1, src2); |
| tcg_gen_add_vec(vece, dst, dst, t); |
| } |
| |
| static void gen_op_fmean16(unsigned vece, uint32_t dofs, uint32_t aofs, |
| uint32_t bofs, uint32_t oprsz, uint32_t maxsz) |
| { |
| static const TCGOpcode vecop_list[] = { |
| INDEX_op_add_vec, INDEX_op_sari_vec, |
| }; |
| static const GVecGen3 op = { |
| .fni8 = gen_helper_fmean16, |
| .fniv = gen_vec_fmean16, |
| .opt_opc = vecop_list, |
| .vece = MO_16, |
| }; |
| tcg_gen_gvec_3(dofs, aofs, bofs, oprsz, maxsz, &op); |
| } |
| #else |
| #define gen_op_fchksm16 ({ qemu_build_not_reached(); NULL; }) |
| #define gen_op_fmean16 ({ qemu_build_not_reached(); NULL; }) |
| #endif |
| |
| static void finishing_insn(DisasContext *dc) |
| { |
| /* |
| * From here, there is no future path through an unwinding exception. |
| * If the current insn cannot raise an exception, the computation of |
| * cpu_cond may be able to be elided. |
| */ |
| if (dc->cpu_cond_live) { |
| tcg_gen_discard_tl(cpu_cond); |
| dc->cpu_cond_live = false; |
| } |
| } |
| |
| static void gen_generic_branch(DisasContext *dc) |
| { |
| TCGv npc0 = tcg_constant_tl(dc->jump_pc[0]); |
| TCGv npc1 = tcg_constant_tl(dc->jump_pc[1]); |
| TCGv c2 = tcg_constant_tl(dc->jump.c2); |
| |
| tcg_gen_movcond_tl(dc->jump.cond, cpu_npc, dc->jump.c1, c2, npc0, npc1); |
| } |
| |
| /* call this function before using the condition register as it may |
| have been set for a jump */ |
| static void flush_cond(DisasContext *dc) |
| { |
| if (dc->npc == JUMP_PC) { |
| gen_generic_branch(dc); |
| dc->npc = DYNAMIC_PC_LOOKUP; |
| } |
| } |
| |
| static void save_npc(DisasContext *dc) |
| { |
| if (dc->npc & 3) { |
| switch (dc->npc) { |
| case JUMP_PC: |
| gen_generic_branch(dc); |
| dc->npc = DYNAMIC_PC_LOOKUP; |
| break; |
| case DYNAMIC_PC: |
| case DYNAMIC_PC_LOOKUP: |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } else { |
| tcg_gen_movi_tl(cpu_npc, dc->npc); |
| } |
| } |
| |
| static void save_state(DisasContext *dc) |
| { |
| tcg_gen_movi_tl(cpu_pc, dc->pc); |
| save_npc(dc); |
| } |
| |
| static void gen_exception(DisasContext *dc, int which) |
| { |
| finishing_insn(dc); |
| save_state(dc); |
| gen_helper_raise_exception(tcg_env, tcg_constant_i32(which)); |
| dc->base.is_jmp = DISAS_NORETURN; |
| } |
| |
| static TCGLabel *delay_exceptionv(DisasContext *dc, TCGv_i32 excp) |
| { |
| DisasDelayException *e = g_new0(DisasDelayException, 1); |
| |
| e->next = dc->delay_excp_list; |
| dc->delay_excp_list = e; |
| |
| e->lab = gen_new_label(); |
| e->excp = excp; |
| e->pc = dc->pc; |
| /* Caller must have used flush_cond before branch. */ |
| assert(e->npc != JUMP_PC); |
| e->npc = dc->npc; |
| |
| return e->lab; |
| } |
| |
| static TCGLabel *delay_exception(DisasContext *dc, int excp) |
| { |
| return delay_exceptionv(dc, tcg_constant_i32(excp)); |
| } |
| |
| static void gen_check_align(DisasContext *dc, TCGv addr, int mask) |
| { |
| TCGv t = tcg_temp_new(); |
| TCGLabel *lab; |
| |
| tcg_gen_andi_tl(t, addr, mask); |
| |
| flush_cond(dc); |
| lab = delay_exception(dc, TT_UNALIGNED); |
| tcg_gen_brcondi_tl(TCG_COND_NE, t, 0, lab); |
| } |
| |
| static void gen_mov_pc_npc(DisasContext *dc) |
| { |
| finishing_insn(dc); |
| |
| if (dc->npc & 3) { |
| switch (dc->npc) { |
| case JUMP_PC: |
| gen_generic_branch(dc); |
| tcg_gen_mov_tl(cpu_pc, cpu_npc); |
| dc->pc = DYNAMIC_PC_LOOKUP; |
| break; |
| case DYNAMIC_PC: |
| case DYNAMIC_PC_LOOKUP: |
| tcg_gen_mov_tl(cpu_pc, cpu_npc); |
| dc->pc = dc->npc; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } else { |
| dc->pc = dc->npc; |
| } |
| } |
| |
| static void gen_compare(DisasCompare *cmp, bool xcc, unsigned int cond, |
| DisasContext *dc) |
| { |
| TCGv t1; |
| |
| cmp->c1 = t1 = tcg_temp_new(); |
| cmp->c2 = 0; |
| |
| switch (cond & 7) { |
| case 0x0: /* never */ |
| cmp->cond = TCG_COND_NEVER; |
| cmp->c1 = tcg_constant_tl(0); |
| break; |
| |
| case 0x1: /* eq: Z */ |
| cmp->cond = TCG_COND_EQ; |
| if (TARGET_LONG_BITS == 32 || xcc) { |
| tcg_gen_mov_tl(t1, cpu_cc_Z); |
| } else { |
| tcg_gen_ext32u_tl(t1, cpu_icc_Z); |
| } |
| break; |
| |
| case 0x2: /* le: Z | (N ^ V) */ |
| /* |
| * Simplify: |
| * cc_Z || (N ^ V) < 0 NE |
| * cc_Z && !((N ^ V) < 0) EQ |
| * cc_Z & ~((N ^ V) >> TLB) EQ |
| */ |
| cmp->cond = TCG_COND_EQ; |
| tcg_gen_xor_tl(t1, cpu_cc_N, cpu_cc_V); |
| tcg_gen_sextract_tl(t1, t1, xcc ? 63 : 31, 1); |
| tcg_gen_andc_tl(t1, xcc ? cpu_cc_Z : cpu_icc_Z, t1); |
| if (TARGET_LONG_BITS == 64 && !xcc) { |
| tcg_gen_ext32u_tl(t1, t1); |
| } |
| break; |
| |
| case 0x3: /* lt: N ^ V */ |
| cmp->cond = TCG_COND_LT; |
| tcg_gen_xor_tl(t1, cpu_cc_N, cpu_cc_V); |
| if (TARGET_LONG_BITS == 64 && !xcc) { |
| tcg_gen_ext32s_tl(t1, t1); |
| } |
| break; |
| |
| case 0x4: /* leu: Z | C */ |
| /* |
| * Simplify: |
| * cc_Z == 0 || cc_C != 0 NE |
| * cc_Z != 0 && cc_C == 0 EQ |
| * cc_Z & (cc_C ? 0 : -1) EQ |
| * cc_Z & (cc_C - 1) EQ |
| */ |
| cmp->cond = TCG_COND_EQ; |
| if (TARGET_LONG_BITS == 32 || xcc) { |
| tcg_gen_subi_tl(t1, cpu_cc_C, 1); |
| tcg_gen_and_tl(t1, t1, cpu_cc_Z); |
| } else { |
| tcg_gen_extract_tl(t1, cpu_icc_C, 32, 1); |
| tcg_gen_subi_tl(t1, t1, 1); |
| tcg_gen_and_tl(t1, t1, cpu_icc_Z); |
| tcg_gen_ext32u_tl(t1, t1); |
| } |
| break; |
| |
| case 0x5: /* ltu: C */ |
| cmp->cond = TCG_COND_NE; |
| if (TARGET_LONG_BITS == 32 || xcc) { |
| tcg_gen_mov_tl(t1, cpu_cc_C); |
| } else { |
| tcg_gen_extract_tl(t1, cpu_icc_C, 32, 1); |
| } |
| break; |
| |
| case 0x6: /* neg: N */ |
| cmp->cond = TCG_COND_LT; |
| if (TARGET_LONG_BITS == 32 || xcc) { |
| tcg_gen_mov_tl(t1, cpu_cc_N); |
| } else { |
| tcg_gen_ext32s_tl(t1, cpu_cc_N); |
| } |
| break; |
| |
| case 0x7: /* vs: V */ |
| cmp->cond = TCG_COND_LT; |
| if (TARGET_LONG_BITS == 32 || xcc) { |
| tcg_gen_mov_tl(t1, cpu_cc_V); |
| } else { |
| tcg_gen_ext32s_tl(t1, cpu_cc_V); |
| } |
| break; |
| } |
| if (cond & 8) { |
| cmp->cond = tcg_invert_cond(cmp->cond); |
| } |
| } |
| |
| static void gen_fcompare(DisasCompare *cmp, unsigned int cc, unsigned int cond) |
| { |
| TCGv_i32 fcc = cpu_fcc[cc]; |
| TCGv_i32 c1 = fcc; |
| int c2 = 0; |
| TCGCond tcond; |
| |
| /* |
| * FCC values: |
| * 0 = |
| * 1 < |
| * 2 > |
| * 3 unordered |
| */ |
| switch (cond & 7) { |
| case 0x0: /* fbn */ |
| tcond = TCG_COND_NEVER; |
| break; |
| case 0x1: /* fbne : !0 */ |
| tcond = TCG_COND_NE; |
| break; |
| case 0x2: /* fblg : 1 or 2 */ |
| /* fcc in {1,2} - 1 -> fcc in {0,1} */ |
| c1 = tcg_temp_new_i32(); |
| tcg_gen_addi_i32(c1, fcc, -1); |
| c2 = 1; |
| tcond = TCG_COND_LEU; |
| break; |
| case 0x3: /* fbul : 1 or 3 */ |
| c1 = tcg_temp_new_i32(); |
| tcg_gen_andi_i32(c1, fcc, 1); |
| tcond = TCG_COND_NE; |
| break; |
| case 0x4: /* fbl : 1 */ |
| c2 = 1; |
| tcond = TCG_COND_EQ; |
| break; |
| case 0x5: /* fbug : 2 or 3 */ |
| c2 = 2; |
| tcond = TCG_COND_GEU; |
| break; |
| case 0x6: /* fbg : 2 */ |
| c2 = 2; |
| tcond = TCG_COND_EQ; |
| break; |
| case 0x7: /* fbu : 3 */ |
| c2 = 3; |
| tcond = TCG_COND_EQ; |
| break; |
| } |
| if (cond & 8) { |
| tcond = tcg_invert_cond(tcond); |
| } |
| |
| cmp->cond = tcond; |
| cmp->c2 = c2; |
| cmp->c1 = tcg_temp_new(); |
| tcg_gen_extu_i32_tl(cmp->c1, c1); |
| } |
| |
| static bool gen_compare_reg(DisasCompare *cmp, int cond, TCGv r_src) |
| { |
| static const TCGCond cond_reg[4] = { |
| TCG_COND_NEVER, /* reserved */ |
| TCG_COND_EQ, |
| TCG_COND_LE, |
| TCG_COND_LT, |
| }; |
| TCGCond tcond; |
| |
| if ((cond & 3) == 0) { |
| return false; |
| } |
| tcond = cond_reg[cond & 3]; |
| if (cond & 4) { |
| tcond = tcg_invert_cond(tcond); |
| } |
| |
| cmp->cond = tcond; |
| cmp->c1 = tcg_temp_new(); |
| cmp->c2 = 0; |
| tcg_gen_mov_tl(cmp->c1, r_src); |
| return true; |
| } |
| |
| static void gen_op_clear_ieee_excp_and_FTT(void) |
| { |
| tcg_gen_st_i32(tcg_constant_i32(0), tcg_env, |
| offsetof(CPUSPARCState, fsr_cexc_ftt)); |
| } |
| |
| static void gen_op_fmovs(TCGv_i32 dst, TCGv_i32 src) |
| { |
| gen_op_clear_ieee_excp_and_FTT(); |
| tcg_gen_mov_i32(dst, src); |
| } |
| |
| static void gen_op_fnegs(TCGv_i32 dst, TCGv_i32 src) |
| { |
| gen_op_clear_ieee_excp_and_FTT(); |
| tcg_gen_xori_i32(dst, src, 1u << 31); |
| } |
| |
| static void gen_op_fabss(TCGv_i32 dst, TCGv_i32 src) |
| { |
| gen_op_clear_ieee_excp_and_FTT(); |
| tcg_gen_andi_i32(dst, src, ~(1u << 31)); |
| } |
| |
| static void gen_op_fmovd(TCGv_i64 dst, TCGv_i64 src) |
| { |
| gen_op_clear_ieee_excp_and_FTT(); |
| tcg_gen_mov_i64(dst, src); |
| } |
| |
| static void gen_op_fnegd(TCGv_i64 dst, TCGv_i64 src) |
| { |
| gen_op_clear_ieee_excp_and_FTT(); |
| tcg_gen_xori_i64(dst, src, 1ull << 63); |
| } |
| |
| static void gen_op_fabsd(TCGv_i64 dst, TCGv_i64 src) |
| { |
| gen_op_clear_ieee_excp_and_FTT(); |
| tcg_gen_andi_i64(dst, src, ~(1ull << 63)); |
| } |
| |
| static void gen_op_fnegq(TCGv_i128 dst, TCGv_i128 src) |
| { |
| TCGv_i64 l = tcg_temp_new_i64(); |
| TCGv_i64 h = tcg_temp_new_i64(); |
| |
| tcg_gen_extr_i128_i64(l, h, src); |
| tcg_gen_xori_i64(h, h, 1ull << 63); |
| tcg_gen_concat_i64_i128(dst, l, h); |
| } |
| |
| static void gen_op_fabsq(TCGv_i128 dst, TCGv_i128 src) |
| { |
| TCGv_i64 l = tcg_temp_new_i64(); |
| TCGv_i64 h = tcg_temp_new_i64(); |
| |
| tcg_gen_extr_i128_i64(l, h, src); |
| tcg_gen_andi_i64(h, h, ~(1ull << 63)); |
| tcg_gen_concat_i64_i128(dst, l, h); |
| } |
| |
| static void gen_op_fmadds(TCGv_i32 d, TCGv_i32 s1, TCGv_i32 s2, TCGv_i32 s3) |
| { |
| gen_helper_fmadds(d, tcg_env, s1, s2, s3, tcg_constant_i32(0)); |
| } |
| |
| static void gen_op_fmaddd(TCGv_i64 d, TCGv_i64 s1, TCGv_i64 s2, TCGv_i64 s3) |
| { |
| gen_helper_fmaddd(d, tcg_env, s1, s2, s3, tcg_constant_i32(0)); |
| } |
| |
| static void gen_op_fmsubs(TCGv_i32 d, TCGv_i32 s1, TCGv_i32 s2, TCGv_i32 s3) |
| { |
| int op = float_muladd_negate_c; |
| gen_helper_fmadds(d, tcg_env, s1, s2, s3, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fmsubd(TCGv_i64 d, TCGv_i64 s1, TCGv_i64 s2, TCGv_i64 s3) |
| { |
| int op = float_muladd_negate_c; |
| gen_helper_fmaddd(d, tcg_env, s1, s2, s3, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fnmsubs(TCGv_i32 d, TCGv_i32 s1, TCGv_i32 s2, TCGv_i32 s3) |
| { |
| int op = float_muladd_negate_c | float_muladd_negate_result; |
| gen_helper_fmadds(d, tcg_env, s1, s2, s3, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fnmsubd(TCGv_i64 d, TCGv_i64 s1, TCGv_i64 s2, TCGv_i64 s3) |
| { |
| int op = float_muladd_negate_c | float_muladd_negate_result; |
| gen_helper_fmaddd(d, tcg_env, s1, s2, s3, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fnmadds(TCGv_i32 d, TCGv_i32 s1, TCGv_i32 s2, TCGv_i32 s3) |
| { |
| int op = float_muladd_negate_result; |
| gen_helper_fmadds(d, tcg_env, s1, s2, s3, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fnmaddd(TCGv_i64 d, TCGv_i64 s1, TCGv_i64 s2, TCGv_i64 s3) |
| { |
| int op = float_muladd_negate_result; |
| gen_helper_fmaddd(d, tcg_env, s1, s2, s3, tcg_constant_i32(op)); |
| } |
| |
| /* Use muladd to compute (1 * src1) + src2 / 2 with one rounding. */ |
| static void gen_op_fhadds(TCGv_i32 d, TCGv_i32 s1, TCGv_i32 s2) |
| { |
| TCGv_i32 one = tcg_constant_i32(float32_one); |
| int op = float_muladd_halve_result; |
| gen_helper_fmadds(d, tcg_env, one, s1, s2, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fhaddd(TCGv_i64 d, TCGv_i64 s1, TCGv_i64 s2) |
| { |
| TCGv_i64 one = tcg_constant_i64(float64_one); |
| int op = float_muladd_halve_result; |
| gen_helper_fmaddd(d, tcg_env, one, s1, s2, tcg_constant_i32(op)); |
| } |
| |
| /* Use muladd to compute (1 * src1) - src2 / 2 with one rounding. */ |
| static void gen_op_fhsubs(TCGv_i32 d, TCGv_i32 s1, TCGv_i32 s2) |
| { |
| TCGv_i32 one = tcg_constant_i32(float32_one); |
| int op = float_muladd_negate_c | float_muladd_halve_result; |
| gen_helper_fmadds(d, tcg_env, one, s1, s2, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fhsubd(TCGv_i64 d, TCGv_i64 s1, TCGv_i64 s2) |
| { |
| TCGv_i64 one = tcg_constant_i64(float64_one); |
| int op = float_muladd_negate_c | float_muladd_halve_result; |
| gen_helper_fmaddd(d, tcg_env, one, s1, s2, tcg_constant_i32(op)); |
| } |
| |
| /* Use muladd to compute -((1 * src1) + src2 / 2) with one rounding. */ |
| static void gen_op_fnhadds(TCGv_i32 d, TCGv_i32 s1, TCGv_i32 s2) |
| { |
| TCGv_i32 one = tcg_constant_i32(float32_one); |
| int op = float_muladd_negate_result | float_muladd_halve_result; |
| gen_helper_fmadds(d, tcg_env, one, s1, s2, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fnhaddd(TCGv_i64 d, TCGv_i64 s1, TCGv_i64 s2) |
| { |
| TCGv_i64 one = tcg_constant_i64(float64_one); |
| int op = float_muladd_negate_result | float_muladd_halve_result; |
| gen_helper_fmaddd(d, tcg_env, one, s1, s2, tcg_constant_i32(op)); |
| } |
| |
| static void gen_op_fpexception_im(DisasContext *dc, int ftt) |
| { |
| /* |
| * CEXC is only set when succesfully completing an FPop, |
| * or when raising FSR_FTT_IEEE_EXCP, i.e. check_ieee_exception. |
| * Thus we can simply store FTT into this field. |
| */ |
| tcg_gen_st_i32(tcg_constant_i32(ftt), tcg_env, |
| offsetof(CPUSPARCState, fsr_cexc_ftt)); |
| gen_exception(dc, TT_FP_EXCP); |
| } |
| |
| static int gen_trap_ifnofpu(DisasContext *dc) |
| { |
| #if !defined(CONFIG_USER_ONLY) |
| if (!dc->fpu_enabled) { |
| gen_exception(dc, TT_NFPU_INSN); |
| return 1; |
| } |
| #endif |
| return 0; |
| } |
| |
| /* asi moves */ |
| typedef enum { |
| GET_ASI_HELPER, |
| GET_ASI_EXCP, |
| GET_ASI_DIRECT, |
| GET_ASI_DTWINX, |
| GET_ASI_CODE, |
| GET_ASI_BLOCK, |
| GET_ASI_SHORT, |
| GET_ASI_BCOPY, |
| GET_ASI_BFILL, |
| } ASIType; |
| |
| typedef struct { |
| ASIType type; |
| int asi; |
| int mem_idx; |
| MemOp memop; |
| } DisasASI; |
| |
| /* |
| * Build DisasASI. |
| * For asi == -1, treat as non-asi. |
| * For ask == -2, treat as immediate offset (v8 error, v9 %asi). |
| */ |
| static DisasASI resolve_asi(DisasContext *dc, int asi, MemOp memop) |
| { |
| ASIType type = GET_ASI_HELPER; |
| int mem_idx = dc->mem_idx; |
| |
| if (asi == -1) { |
| /* Artificial "non-asi" case. */ |
| type = GET_ASI_DIRECT; |
| goto done; |
| } |
| |
| #ifndef TARGET_SPARC64 |
| /* Before v9, all asis are immediate and privileged. */ |
| if (asi < 0) { |
| gen_exception(dc, TT_ILL_INSN); |
| type = GET_ASI_EXCP; |
| } else if (supervisor(dc) |
| /* Note that LEON accepts ASI_USERDATA in user mode, for |
| use with CASA. Also note that previous versions of |
| QEMU allowed (and old versions of gcc emitted) ASI_P |
| for LEON, which is incorrect. */ |
| || (asi == ASI_USERDATA |
| && (dc->def->features & CPU_FEATURE_CASA))) { |
| switch (asi) { |
| case ASI_USERDATA: /* User data access */ |
| mem_idx = MMU_USER_IDX; |
| type = GET_ASI_DIRECT; |
| break; |
| case ASI_KERNELDATA: /* Supervisor data access */ |
| mem_idx = MMU_KERNEL_IDX; |
| type = GET_ASI_DIRECT; |
| break; |
| case ASI_USERTXT: /* User text access */ |
| mem_idx = MMU_USER_IDX; |
| type = GET_ASI_CODE; |
| break; |
| case ASI_KERNELTXT: /* Supervisor text access */ |
| mem_idx = MMU_KERNEL_IDX; |
| type = GET_ASI_CODE; |
| break; |
| case ASI_M_BYPASS: /* MMU passthrough */ |
| case ASI_LEON_BYPASS: /* LEON MMU passthrough */ |
| mem_idx = MMU_PHYS_IDX; |
| type = GET_ASI_DIRECT; |
| break; |
| case ASI_M_BCOPY: /* Block copy, sta access */ |
| mem_idx = MMU_KERNEL_IDX; |
| type = GET_ASI_BCOPY; |
| break; |
| case ASI_M_BFILL: /* Block fill, stda access */ |
| mem_idx = MMU_KERNEL_IDX; |
| type = GET_ASI_BFILL; |
| break; |
| } |
| |
| /* MMU_PHYS_IDX is used when the MMU is disabled to passthrough the |
| * permissions check in get_physical_address(..). |
| */ |
| mem_idx = (dc->mem_idx == MMU_PHYS_IDX) ? MMU_PHYS_IDX : mem_idx; |
| } else { |
| gen_exception(dc, TT_PRIV_INSN); |
| type = GET_ASI_EXCP; |
| } |
| #else |
| if (asi < 0) { |
| asi = dc->asi; |
| } |
| /* With v9, all asis below 0x80 are privileged. */ |
| /* ??? We ought to check cpu_has_hypervisor, but we didn't copy |
| down that bit into DisasContext. For the moment that's ok, |
| since the direct implementations below doesn't have any ASIs |
| in the restricted [0x30, 0x7f] range, and the check will be |
| done properly in the helper. */ |
| if (!supervisor(dc) && asi < 0x80) { |
| gen_exception(dc, TT_PRIV_ACT); |
| type = GET_ASI_EXCP; |
| } else { |
| switch (asi) { |
| case ASI_REAL: /* Bypass */ |
| case ASI_REAL_IO: /* Bypass, non-cacheable */ |
| case ASI_REAL_L: /* Bypass LE */ |
| case ASI_REAL_IO_L: /* Bypass, non-cacheable LE */ |
| case ASI_TWINX_REAL: /* Real address, twinx */ |
| case ASI_TWINX_REAL_L: /* Real address, twinx, LE */ |
| case ASI_QUAD_LDD_PHYS: |
| case ASI_QUAD_LDD_PHYS_L: |
| mem_idx = MMU_PHYS_IDX; |
| break; |
| case ASI_N: /* Nucleus */ |
| case ASI_NL: /* Nucleus LE */ |
| case ASI_TWINX_N: |
| case ASI_TWINX_NL: |
| case ASI_NUCLEUS_QUAD_LDD: |
| case ASI_NUCLEUS_QUAD_LDD_L: |
| if (hypervisor(dc)) { |
| mem_idx = MMU_PHYS_IDX; |
| } else { |
| mem_idx = MMU_NUCLEUS_IDX; |
| } |
| break; |
| case ASI_AIUP: /* As if user primary */ |
| case ASI_AIUPL: /* As if user primary LE */ |
| case ASI_TWINX_AIUP: |
| case ASI_TWINX_AIUP_L: |
| case ASI_BLK_AIUP_4V: |
| case ASI_BLK_AIUP_L_4V: |
| case ASI_BLK_AIUP: |
| case ASI_BLK_AIUPL: |
| case ASI_MON_AIUP: |
| mem_idx = MMU_USER_IDX; |
| break; |
| case ASI_AIUS: /* As if user secondary */ |
| case ASI_AIUSL: /* As if user secondary LE */ |
| case ASI_TWINX_AIUS: |
| case ASI_TWINX_AIUS_L: |
| case ASI_BLK_AIUS_4V: |
| case ASI_BLK_AIUS_L_4V: |
| case ASI_BLK_AIUS: |
| case ASI_BLK_AIUSL: |
| case ASI_MON_AIUS: |
| mem_idx = MMU_USER_SECONDARY_IDX; |
| break; |
| case ASI_S: /* Secondary */ |
| case ASI_SL: /* Secondary LE */ |
| case ASI_TWINX_S: |
| case ASI_TWINX_SL: |
| case ASI_BLK_COMMIT_S: |
| case ASI_BLK_S: |
| case ASI_BLK_SL: |
| case ASI_FL8_S: |
| case ASI_FL8_SL: |
| case ASI_FL16_S: |
| case ASI_FL16_SL: |
| case ASI_MON_S: |
| if (mem_idx == MMU_USER_IDX) { |
| mem_idx = MMU_USER_SECONDARY_IDX; |
| } else if (mem_idx == MMU_KERNEL_IDX) { |
| mem_idx = MMU_KERNEL_SECONDARY_IDX; |
| } |
| break; |
| case ASI_P: /* Primary */ |
| case ASI_PL: /* Primary LE */ |
| case ASI_TWINX_P: |
| case ASI_TWINX_PL: |
| case ASI_BLK_COMMIT_P: |
| case ASI_BLK_P: |
| case ASI_BLK_PL: |
| case ASI_FL8_P: |
| case ASI_FL8_PL: |
| case ASI_FL16_P: |
| case ASI_FL16_PL: |
| case ASI_MON_P: |
| break; |
| } |
| switch (asi) { |
| case ASI_REAL: |
| case ASI_REAL_IO: |
| case ASI_REAL_L: |
| case ASI_REAL_IO_L: |
| case ASI_N: |
| case ASI_NL: |
| case ASI_AIUP: |
| case ASI_AIUPL: |
| case ASI_AIUS: |
| case ASI_AIUSL: |
| case ASI_S: |
| case ASI_SL: |
| case ASI_P: |
| case ASI_PL: |
| case ASI_MON_P: |
| case ASI_MON_S: |
| case ASI_MON_AIUP: |
| case ASI_MON_AIUS: |
| type = GET_ASI_DIRECT; |
| break; |
| case ASI_TWINX_REAL: |
| case ASI_TWINX_REAL_L: |
| case ASI_TWINX_N: |
| case ASI_TWINX_NL: |
| case ASI_TWINX_AIUP: |
| case ASI_TWINX_AIUP_L: |
| case ASI_TWINX_AIUS: |
| case ASI_TWINX_AIUS_L: |
| case ASI_TWINX_P: |
| case ASI_TWINX_PL: |
| case ASI_TWINX_S: |
| case ASI_TWINX_SL: |
| case ASI_QUAD_LDD_PHYS: |
| case ASI_QUAD_LDD_PHYS_L: |
| case ASI_NUCLEUS_QUAD_LDD: |
| case ASI_NUCLEUS_QUAD_LDD_L: |
| type = GET_ASI_DTWINX; |
| break; |
| case ASI_BLK_COMMIT_P: |
| case ASI_BLK_COMMIT_S: |
| case ASI_BLK_AIUP_4V: |
| case ASI_BLK_AIUP_L_4V: |
| case ASI_BLK_AIUP: |
| case ASI_BLK_AIUPL: |
| case ASI_BLK_AIUS_4V: |
| case ASI_BLK_AIUS_L_4V: |
| case ASI_BLK_AIUS: |
| case ASI_BLK_AIUSL: |
| case ASI_BLK_S: |
| case ASI_BLK_SL: |
| case ASI_BLK_P: |
| case ASI_BLK_PL: |
| type = GET_ASI_BLOCK; |
| break; |
| case ASI_FL8_S: |
| case ASI_FL8_SL: |
| case ASI_FL8_P: |
| case ASI_FL8_PL: |
| memop = MO_UB; |
| type = GET_ASI_SHORT; |
| break; |
| case ASI_FL16_S: |
| case ASI_FL16_SL: |
| case ASI_FL16_P: |
| case ASI_FL16_PL: |
| memop = MO_TEUW; |
| type = GET_ASI_SHORT; |
| break; |
| } |
| /* The little-endian asis all have bit 3 set. */ |
| if (asi & 8) { |
| memop ^= MO_BSWAP; |
| } |
| } |
| #endif |
| |
| done: |
| return (DisasASI){ type, asi, mem_idx, memop }; |
| } |
| |
| #if defined(CONFIG_USER_ONLY) && !defined(TARGET_SPARC64) |
| static void gen_helper_ld_asi(TCGv_i64 r, TCGv_env e, TCGv a, |
| TCGv_i32 asi, TCGv_i32 mop) |
| { |
| g_assert_not_reached(); |
| } |
| |
| static void gen_helper_st_asi(TCGv_env e, TCGv a, TCGv_i64 r, |
| TCGv_i32 asi, TCGv_i32 mop) |
| { |
| g_assert_not_reached(); |
| } |
| #endif |
| |
| static void gen_ld_asi(DisasContext *dc, DisasASI *da, TCGv dst, TCGv addr) |
| { |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| break; |
| case GET_ASI_DTWINX: /* Reserved for ldda. */ |
| gen_exception(dc, TT_ILL_INSN); |
| break; |
| case GET_ASI_DIRECT: |
| tcg_gen_qemu_ld_tl(dst, addr, da->mem_idx, da->memop | MO_ALIGN); |
| break; |
| |
| case GET_ASI_CODE: |
| #if !defined(CONFIG_USER_ONLY) && !defined(TARGET_SPARC64) |
| { |
| MemOpIdx oi = make_memop_idx(da->memop, da->mem_idx); |
| TCGv_i64 t64 = tcg_temp_new_i64(); |
| |
| gen_helper_ld_code(t64, tcg_env, addr, tcg_constant_i32(oi)); |
| tcg_gen_trunc_i64_tl(dst, t64); |
| } |
| break; |
| #else |
| g_assert_not_reached(); |
| #endif |
| |
| default: |
| { |
| TCGv_i32 r_asi = tcg_constant_i32(da->asi); |
| TCGv_i32 r_mop = tcg_constant_i32(da->memop | MO_ALIGN); |
| |
| save_state(dc); |
| #ifdef TARGET_SPARC64 |
| gen_helper_ld_asi(dst, tcg_env, addr, r_asi, r_mop); |
| #else |
| { |
| TCGv_i64 t64 = tcg_temp_new_i64(); |
| gen_helper_ld_asi(t64, tcg_env, addr, r_asi, r_mop); |
| tcg_gen_trunc_i64_tl(dst, t64); |
| } |
| #endif |
| } |
| break; |
| } |
| } |
| |
| static void gen_st_asi(DisasContext *dc, DisasASI *da, TCGv src, TCGv addr) |
| { |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| break; |
| |
| case GET_ASI_DTWINX: /* Reserved for stda. */ |
| if (TARGET_LONG_BITS == 32) { |
| gen_exception(dc, TT_ILL_INSN); |
| break; |
| } else if (!(dc->def->features & CPU_FEATURE_HYPV)) { |
| /* Pre OpenSPARC CPUs don't have these */ |
| gen_exception(dc, TT_ILL_INSN); |
| break; |
| } |
| /* In OpenSPARC T1+ CPUs TWINX ASIs in store are ST_BLKINIT_ ASIs */ |
| /* fall through */ |
| |
| case GET_ASI_DIRECT: |
| tcg_gen_qemu_st_tl(src, addr, da->mem_idx, da->memop | MO_ALIGN); |
| break; |
| |
| case GET_ASI_BCOPY: |
| assert(TARGET_LONG_BITS == 32); |
| /* |
| * Copy 32 bytes from the address in SRC to ADDR. |
| * |
| * From Ross RT625 hyperSPARC manual, section 4.6: |
| * "Block Copy and Block Fill will work only on cache line boundaries." |
| * |
| * It does not specify if an unaliged address is truncated or trapped. |
| * Previous qemu behaviour was to truncate to 4 byte alignment, which |
| * is obviously wrong. The only place I can see this used is in the |
| * Linux kernel which begins with page alignment, advancing by 32, |
| * so is always aligned. Assume truncation as the simpler option. |
| * |
| * Since the loads and stores are paired, allow the copy to happen |
| * in the host endianness. The copy need not be atomic. |
| */ |
| { |
| MemOp mop = MO_128 | MO_ATOM_IFALIGN_PAIR; |
| TCGv saddr = tcg_temp_new(); |
| TCGv daddr = tcg_temp_new(); |
| TCGv_i128 tmp = tcg_temp_new_i128(); |
| |
| tcg_gen_andi_tl(saddr, src, -32); |
| tcg_gen_andi_tl(daddr, addr, -32); |
| tcg_gen_qemu_ld_i128(tmp, saddr, da->mem_idx, mop); |
| tcg_gen_qemu_st_i128(tmp, daddr, da->mem_idx, mop); |
| tcg_gen_addi_tl(saddr, saddr, 16); |
| tcg_gen_addi_tl(daddr, daddr, 16); |
| tcg_gen_qemu_ld_i128(tmp, saddr, da->mem_idx, mop); |
| tcg_gen_qemu_st_i128(tmp, daddr, da->mem_idx, mop); |
| } |
| break; |
| |
| default: |
| { |
| TCGv_i32 r_asi = tcg_constant_i32(da->asi); |
| TCGv_i32 r_mop = tcg_constant_i32(da->memop | MO_ALIGN); |
| |
| save_state(dc); |
| #ifdef TARGET_SPARC64 |
| gen_helper_st_asi(tcg_env, addr, src, r_asi, r_mop); |
| #else |
| { |
| TCGv_i64 t64 = tcg_temp_new_i64(); |
| tcg_gen_extu_tl_i64(t64, src); |
| gen_helper_st_asi(tcg_env, addr, t64, r_asi, r_mop); |
| } |
| #endif |
| |
| /* A write to a TLB register may alter page maps. End the TB. */ |
| dc->npc = DYNAMIC_PC; |
| } |
| break; |
| } |
| } |
| |
| static void gen_swap_asi(DisasContext *dc, DisasASI *da, |
| TCGv dst, TCGv src, TCGv addr) |
| { |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| break; |
| case GET_ASI_DIRECT: |
| tcg_gen_atomic_xchg_tl(dst, addr, src, |
| da->mem_idx, da->memop | MO_ALIGN); |
| break; |
| default: |
| /* ??? Should be DAE_invalid_asi. */ |
| gen_exception(dc, TT_DATA_ACCESS); |
| break; |
| } |
| } |
| |
| static void gen_cas_asi(DisasContext *dc, DisasASI *da, |
| TCGv oldv, TCGv newv, TCGv cmpv, TCGv addr) |
| { |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| return; |
| case GET_ASI_DIRECT: |
| tcg_gen_atomic_cmpxchg_tl(oldv, addr, cmpv, newv, |
| da->mem_idx, da->memop | MO_ALIGN); |
| break; |
| default: |
| /* ??? Should be DAE_invalid_asi. */ |
| gen_exception(dc, TT_DATA_ACCESS); |
| break; |
| } |
| } |
| |
| static void gen_ldstub_asi(DisasContext *dc, DisasASI *da, TCGv dst, TCGv addr) |
| { |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| break; |
| case GET_ASI_DIRECT: |
| tcg_gen_atomic_xchg_tl(dst, addr, tcg_constant_tl(0xff), |
| da->mem_idx, MO_UB); |
| break; |
| default: |
| /* ??? In theory, this should be raise DAE_invalid_asi. |
| But the SS-20 roms do ldstuba [%l0] #ASI_M_CTL, %o1. */ |
| if (tb_cflags(dc->base.tb) & CF_PARALLEL) { |
| gen_helper_exit_atomic(tcg_env); |
| } else { |
| TCGv_i32 r_asi = tcg_constant_i32(da->asi); |
| TCGv_i32 r_mop = tcg_constant_i32(MO_UB); |
| TCGv_i64 s64, t64; |
| |
| save_state(dc); |
| t64 = tcg_temp_new_i64(); |
| gen_helper_ld_asi(t64, tcg_env, addr, r_asi, r_mop); |
| |
| s64 = tcg_constant_i64(0xff); |
| gen_helper_st_asi(tcg_env, addr, s64, r_asi, r_mop); |
| |
| tcg_gen_trunc_i64_tl(dst, t64); |
| |
| /* End the TB. */ |
| dc->npc = DYNAMIC_PC; |
| } |
| break; |
| } |
| } |
| |
| static void gen_ldf_asi(DisasContext *dc, DisasASI *da, MemOp orig_size, |
| TCGv addr, int rd) |
| { |
| MemOp memop = da->memop; |
| MemOp size = memop & MO_SIZE; |
| TCGv_i32 d32; |
| TCGv_i64 d64, l64; |
| TCGv addr_tmp; |
| |
| /* TODO: Use 128-bit load/store below. */ |
| if (size == MO_128) { |
| memop = (memop & ~MO_SIZE) | MO_64; |
| } |
| |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| break; |
| |
| case GET_ASI_DIRECT: |
| memop |= MO_ALIGN_4; |
| switch (size) { |
| case MO_32: |
| d32 = tcg_temp_new_i32(); |
| tcg_gen_qemu_ld_i32(d32, addr, da->mem_idx, memop); |
| gen_store_fpr_F(dc, rd, d32); |
| break; |
| |
| case MO_64: |
| d64 = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(d64, addr, da->mem_idx, memop); |
| gen_store_fpr_D(dc, rd, d64); |
| break; |
| |
| case MO_128: |
| d64 = tcg_temp_new_i64(); |
| l64 = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(d64, addr, da->mem_idx, memop); |
| addr_tmp = tcg_temp_new(); |
| tcg_gen_addi_tl(addr_tmp, addr, 8); |
| tcg_gen_qemu_ld_i64(l64, addr_tmp, da->mem_idx, memop); |
| gen_store_fpr_D(dc, rd, d64); |
| gen_store_fpr_D(dc, rd + 2, l64); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| break; |
| |
| case GET_ASI_BLOCK: |
| /* Valid for lddfa on aligned registers only. */ |
| if (orig_size == MO_64 && (rd & 7) == 0) { |
| /* The first operation checks required alignment. */ |
| addr_tmp = tcg_temp_new(); |
| d64 = tcg_temp_new_i64(); |
| for (int i = 0; ; ++i) { |
| tcg_gen_qemu_ld_i64(d64, addr, da->mem_idx, |
| memop | (i == 0 ? MO_ALIGN_64 : 0)); |
| gen_store_fpr_D(dc, rd + 2 * i, d64); |
| if (i == 7) { |
| break; |
| } |
| tcg_gen_addi_tl(addr_tmp, addr, 8); |
| addr = addr_tmp; |
| } |
| } else { |
| gen_exception(dc, TT_ILL_INSN); |
| } |
| break; |
| |
| case GET_ASI_SHORT: |
| /* Valid for lddfa only. */ |
| if (orig_size == MO_64) { |
| d64 = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(d64, addr, da->mem_idx, memop | MO_ALIGN); |
| gen_store_fpr_D(dc, rd, d64); |
| } else { |
| gen_exception(dc, TT_ILL_INSN); |
| } |
| break; |
| |
| default: |
| { |
| TCGv_i32 r_asi = tcg_constant_i32(da->asi); |
| TCGv_i32 r_mop = tcg_constant_i32(memop | MO_ALIGN); |
| |
| save_state(dc); |
| /* According to the table in the UA2011 manual, the only |
| other asis that are valid for ldfa/lddfa/ldqfa are |
| the NO_FAULT asis. We still need a helper for these, |
| but we can just use the integer asi helper for them. */ |
| switch (size) { |
| case MO_32: |
| d64 = tcg_temp_new_i64(); |
| gen_helper_ld_asi(d64, tcg_env, addr, r_asi, r_mop); |
| d32 = tcg_temp_new_i32(); |
| tcg_gen_extrl_i64_i32(d32, d64); |
| gen_store_fpr_F(dc, rd, d32); |
| break; |
| case MO_64: |
| d64 = tcg_temp_new_i64(); |
| gen_helper_ld_asi(d64, tcg_env, addr, r_asi, r_mop); |
| gen_store_fpr_D(dc, rd, d64); |
| break; |
| case MO_128: |
| d64 = tcg_temp_new_i64(); |
| l64 = tcg_temp_new_i64(); |
| gen_helper_ld_asi(d64, tcg_env, addr, r_asi, r_mop); |
| addr_tmp = tcg_temp_new(); |
| tcg_gen_addi_tl(addr_tmp, addr, 8); |
| gen_helper_ld_asi(l64, tcg_env, addr_tmp, r_asi, r_mop); |
| gen_store_fpr_D(dc, rd, d64); |
| gen_store_fpr_D(dc, rd + 2, l64); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| break; |
| } |
| } |
| |
| static void gen_stf_asi(DisasContext *dc, DisasASI *da, MemOp orig_size, |
| TCGv addr, int rd) |
| { |
| MemOp memop = da->memop; |
| MemOp size = memop & MO_SIZE; |
| TCGv_i32 d32; |
| TCGv_i64 d64; |
| TCGv addr_tmp; |
| |
| /* TODO: Use 128-bit load/store below. */ |
| if (size == MO_128) { |
| memop = (memop & ~MO_SIZE) | MO_64; |
| } |
| |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| break; |
| |
| case GET_ASI_DIRECT: |
| memop |= MO_ALIGN_4; |
| switch (size) { |
| case MO_32: |
| d32 = gen_load_fpr_F(dc, rd); |
| tcg_gen_qemu_st_i32(d32, addr, da->mem_idx, memop | MO_ALIGN); |
| break; |
| case MO_64: |
| d64 = gen_load_fpr_D(dc, rd); |
| tcg_gen_qemu_st_i64(d64, addr, da->mem_idx, memop | MO_ALIGN_4); |
| break; |
| case MO_128: |
| /* Only 4-byte alignment required. However, it is legal for the |
| cpu to signal the alignment fault, and the OS trap handler is |
| required to fix it up. Requiring 16-byte alignment here avoids |
| having to probe the second page before performing the first |
| write. */ |
| d64 = gen_load_fpr_D(dc, rd); |
| tcg_gen_qemu_st_i64(d64, addr, da->mem_idx, memop | MO_ALIGN_16); |
| addr_tmp = tcg_temp_new(); |
| tcg_gen_addi_tl(addr_tmp, addr, 8); |
| d64 = gen_load_fpr_D(dc, rd + 2); |
| tcg_gen_qemu_st_i64(d64, addr_tmp, da->mem_idx, memop); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| break; |
| |
| case GET_ASI_BLOCK: |
| /* Valid for stdfa on aligned registers only. */ |
| if (orig_size == MO_64 && (rd & 7) == 0) { |
| /* The first operation checks required alignment. */ |
| addr_tmp = tcg_temp_new(); |
| for (int i = 0; ; ++i) { |
| d64 = gen_load_fpr_D(dc, rd + 2 * i); |
| tcg_gen_qemu_st_i64(d64, addr, da->mem_idx, |
| memop | (i == 0 ? MO_ALIGN_64 : 0)); |
| if (i == 7) { |
| break; |
| } |
| tcg_gen_addi_tl(addr_tmp, addr, 8); |
| addr = addr_tmp; |
| } |
| } else { |
| gen_exception(dc, TT_ILL_INSN); |
| } |
| break; |
| |
| case GET_ASI_SHORT: |
| /* Valid for stdfa only. */ |
| if (orig_size == MO_64) { |
| d64 = gen_load_fpr_D(dc, rd); |
| tcg_gen_qemu_st_i64(d64, addr, da->mem_idx, memop | MO_ALIGN); |
| } else { |
| gen_exception(dc, TT_ILL_INSN); |
| } |
| break; |
| |
| default: |
| /* According to the table in the UA2011 manual, the only |
| other asis that are valid for ldfa/lddfa/ldqfa are |
| the PST* asis, which aren't currently handled. */ |
| gen_exception(dc, TT_ILL_INSN); |
| break; |
| } |
| } |
| |
| static void gen_ldda_asi(DisasContext *dc, DisasASI *da, TCGv addr, int rd) |
| { |
| TCGv hi = gen_dest_gpr(dc, rd); |
| TCGv lo = gen_dest_gpr(dc, rd + 1); |
| |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| return; |
| |
| case GET_ASI_DTWINX: |
| #ifdef TARGET_SPARC64 |
| { |
| MemOp mop = (da->memop & MO_BSWAP) | MO_128 | MO_ALIGN_16; |
| TCGv_i128 t = tcg_temp_new_i128(); |
| |
| tcg_gen_qemu_ld_i128(t, addr, da->mem_idx, mop); |
| /* |
| * Note that LE twinx acts as if each 64-bit register result is |
| * byte swapped. We perform one 128-bit LE load, so must swap |
| * the order of the writebacks. |
| */ |
| if ((mop & MO_BSWAP) == MO_TE) { |
| tcg_gen_extr_i128_i64(lo, hi, t); |
| } else { |
| tcg_gen_extr_i128_i64(hi, lo, t); |
| } |
| } |
| break; |
| #else |
| g_assert_not_reached(); |
| #endif |
| |
| case GET_ASI_DIRECT: |
| { |
| TCGv_i64 tmp = tcg_temp_new_i64(); |
| |
| tcg_gen_qemu_ld_i64(tmp, addr, da->mem_idx, da->memop | MO_ALIGN); |
| |
| /* Note that LE ldda acts as if each 32-bit register |
| result is byte swapped. Having just performed one |
| 64-bit bswap, we need now to swap the writebacks. */ |
| if ((da->memop & MO_BSWAP) == MO_TE) { |
| tcg_gen_extr_i64_tl(lo, hi, tmp); |
| } else { |
| tcg_gen_extr_i64_tl(hi, lo, tmp); |
| } |
| } |
| break; |
| |
| case GET_ASI_CODE: |
| #if !defined(CONFIG_USER_ONLY) && !defined(TARGET_SPARC64) |
| { |
| MemOpIdx oi = make_memop_idx(da->memop, da->mem_idx); |
| TCGv_i64 tmp = tcg_temp_new_i64(); |
| |
| gen_helper_ld_code(tmp, tcg_env, addr, tcg_constant_i32(oi)); |
| |
| /* See above. */ |
| if ((da->memop & MO_BSWAP) == MO_TE) { |
| tcg_gen_extr_i64_tl(lo, hi, tmp); |
| } else { |
| tcg_gen_extr_i64_tl(hi, lo, tmp); |
| } |
| } |
| break; |
| #else |
| g_assert_not_reached(); |
| #endif |
| |
| default: |
| /* ??? In theory we've handled all of the ASIs that are valid |
| for ldda, and this should raise DAE_invalid_asi. However, |
| real hardware allows others. This can be seen with e.g. |
| FreeBSD 10.3 wrt ASI_IC_TAG. */ |
| { |
| TCGv_i32 r_asi = tcg_constant_i32(da->asi); |
| TCGv_i32 r_mop = tcg_constant_i32(da->memop); |
| TCGv_i64 tmp = tcg_temp_new_i64(); |
| |
| save_state(dc); |
| gen_helper_ld_asi(tmp, tcg_env, addr, r_asi, r_mop); |
| |
| /* See above. */ |
| if ((da->memop & MO_BSWAP) == MO_TE) { |
| tcg_gen_extr_i64_tl(lo, hi, tmp); |
| } else { |
| tcg_gen_extr_i64_tl(hi, lo, tmp); |
| } |
| } |
| break; |
| } |
| |
| gen_store_gpr(dc, rd, hi); |
| gen_store_gpr(dc, rd + 1, lo); |
| } |
| |
| static void gen_stda_asi(DisasContext *dc, DisasASI *da, TCGv addr, int rd) |
| { |
| TCGv hi = gen_load_gpr(dc, rd); |
| TCGv lo = gen_load_gpr(dc, rd + 1); |
| |
| switch (da->type) { |
| case GET_ASI_EXCP: |
| break; |
| |
| case GET_ASI_DTWINX: |
| #ifdef TARGET_SPARC64 |
| { |
| MemOp mop = (da->memop & MO_BSWAP) | MO_128 | MO_ALIGN_16; |
| TCGv_i128 t = tcg_temp_new_i128(); |
| |
| /* |
| * Note that LE twinx acts as if each 64-bit register result is |
| * byte swapped. We perform one 128-bit LE store, so must swap |
| * the order of the construction. |
| */ |
| if ((mop & MO_BSWAP) == MO_TE) { |
| tcg_gen_concat_i64_i128(t, lo, hi); |
| } else { |
| tcg_gen_concat_i64_i128(t, hi, lo); |
| } |
| tcg_gen_qemu_st_i128(t, addr, da->mem_idx, mop); |
| } |
| break; |
| #else |
| g_assert_not_reached(); |
| #endif |
| |
| case GET_ASI_DIRECT: |
| { |
| TCGv_i64 t64 = tcg_temp_new_i64(); |
| |
| /* Note that LE stda acts as if each 32-bit register result is |
| byte swapped. We will perform one 64-bit LE store, so now |
| we must swap the order of the construction. */ |
| if ((da->memop & MO_BSWAP) == MO_TE) { |
| tcg_gen_concat_tl_i64(t64, lo, hi); |
| } else { |
| tcg_gen_concat_tl_i64(t64, hi, lo); |
| } |
| tcg_gen_qemu_st_i64(t64, addr, da->mem_idx, da->memop | MO_ALIGN); |
| } |
| break; |
| |
| case GET_ASI_BFILL: |
| assert(TARGET_LONG_BITS == 32); |
| /* |
| * Store 32 bytes of [rd:rd+1] to ADDR. |
| * See comments for GET_ASI_COPY above. |
| */ |
| { |
| MemOp mop = MO_TE | MO_128 | MO_ATOM_IFALIGN_PAIR; |
| TCGv_i64 t8 = tcg_temp_new_i64(); |
| TCGv_i128 t16 = tcg_temp_new_i128(); |
| TCGv daddr = tcg_temp_new(); |
| |
| tcg_gen_concat_tl_i64(t8, lo, hi); |
| tcg_gen_concat_i64_i128(t16, t8, t8); |
| tcg_gen_andi_tl(daddr, addr, -32); |
| tcg_gen_qemu_st_i128(t16, daddr, da->mem_idx, mop); |
| tcg_gen_addi_tl(daddr, daddr, 16); |
| tcg_gen_qemu_st_i128(t16, daddr, da->mem_idx, mop); |
| } |
| break; |
| |
| default: |
| /* ??? In theory we've handled all of the ASIs that are valid |
| for stda, and this should raise DAE_invalid_asi. */ |
| { |
| TCGv_i32 r_asi = tcg_constant_i32(da->asi); |
| TCGv_i32 r_mop = tcg_constant_i32(da->memop); |
| TCGv_i64 t64 = tcg_temp_new_i64(); |
| |
| /* See above. */ |
| if ((da->memop & MO_BSWAP) == MO_TE) { |
| tcg_gen_concat_tl_i64(t64, lo, hi); |
| } else { |
| tcg_gen_concat_tl_i64(t64, hi, lo); |
| } |
| |
| save_state(dc); |
| gen_helper_st_asi(tcg_env, addr, t64, r_asi, r_mop); |
| } |
| break; |
| } |
| } |
| |
| static void gen_fmovs(DisasContext *dc, DisasCompare *cmp, int rd, int rs) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_i32 c32, zero, dst, s1, s2; |
| TCGv_i64 c64 = tcg_temp_new_i64(); |
| |
| /* We have two choices here: extend the 32 bit data and use movcond_i64, |
| or fold the comparison down to 32 bits and use movcond_i32. Choose |
| the later. */ |
| c32 = tcg_temp_new_i32(); |
| tcg_gen_setcondi_i64(cmp->cond, c64, cmp->c1, cmp->c2); |
| tcg_gen_extrl_i64_i32(c32, c64); |
| |
| s1 = gen_load_fpr_F(dc, rs); |
| s2 = gen_load_fpr_F(dc, rd); |
| dst = tcg_temp_new_i32(); |
| zero = tcg_constant_i32(0); |
| |
| tcg_gen_movcond_i32(TCG_COND_NE, dst, c32, zero, s1, s2); |
| |
| gen_store_fpr_F(dc, rd, dst); |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| static void gen_fmovd(DisasContext *dc, DisasCompare *cmp, int rd, int rs) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_i64 dst = tcg_temp_new_i64(); |
| tcg_gen_movcond_i64(cmp->cond, dst, cmp->c1, tcg_constant_tl(cmp->c2), |
| gen_load_fpr_D(dc, rs), |
| gen_load_fpr_D(dc, rd)); |
| gen_store_fpr_D(dc, rd, dst); |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| static void gen_fmovq(DisasContext *dc, DisasCompare *cmp, int rd, int rs) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv c2 = tcg_constant_tl(cmp->c2); |
| TCGv_i64 h = tcg_temp_new_i64(); |
| TCGv_i64 l = tcg_temp_new_i64(); |
| |
| tcg_gen_movcond_i64(cmp->cond, h, cmp->c1, c2, |
| gen_load_fpr_D(dc, rs), |
| gen_load_fpr_D(dc, rd)); |
| tcg_gen_movcond_i64(cmp->cond, l, cmp->c1, c2, |
| gen_load_fpr_D(dc, rs + 2), |
| gen_load_fpr_D(dc, rd + 2)); |
| gen_store_fpr_D(dc, rd, h); |
| gen_store_fpr_D(dc, rd + 2, l); |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| #ifdef TARGET_SPARC64 |
| static void gen_load_trap_state_at_tl(TCGv_ptr r_tsptr) |
| { |
| TCGv_i32 r_tl = tcg_temp_new_i32(); |
| |
| /* load env->tl into r_tl */ |
| tcg_gen_ld_i32(r_tl, tcg_env, offsetof(CPUSPARCState, tl)); |
| |
| /* tl = [0 ... MAXTL_MASK] where MAXTL_MASK must be power of 2 */ |
| tcg_gen_andi_i32(r_tl, r_tl, MAXTL_MASK); |
| |
| /* calculate offset to current trap state from env->ts, reuse r_tl */ |
| tcg_gen_muli_i32(r_tl, r_tl, sizeof (trap_state)); |
| tcg_gen_addi_ptr(r_tsptr, tcg_env, offsetof(CPUSPARCState, ts)); |
| |
| /* tsptr = env->ts[env->tl & MAXTL_MASK] */ |
| { |
| TCGv_ptr r_tl_tmp = tcg_temp_new_ptr(); |
| tcg_gen_ext_i32_ptr(r_tl_tmp, r_tl); |
| tcg_gen_add_ptr(r_tsptr, r_tsptr, r_tl_tmp); |
| } |
| } |
| #endif |
| |
| static int extract_dfpreg(DisasContext *dc, int x) |
| { |
| int r = x & 0x1e; |
| #ifdef TARGET_SPARC64 |
| r |= (x & 1) << 5; |
| #endif |
| return r; |
| } |
| |
| static int extract_qfpreg(DisasContext *dc, int x) |
| { |
| int r = x & 0x1c; |
| #ifdef TARGET_SPARC64 |
| r |= (x & 1) << 5; |
| #endif |
| return r; |
| } |
| |
| /* Include the auto-generated decoder. */ |
| #include "decode-insns.c.inc" |
| |
| #define TRANS(NAME, AVAIL, FUNC, ...) \ |
| static bool trans_##NAME(DisasContext *dc, arg_##NAME *a) \ |
| { return avail_##AVAIL(dc) && FUNC(dc, __VA_ARGS__); } |
| |
| #define avail_ALL(C) true |
| #ifdef TARGET_SPARC64 |
| # define avail_32(C) false |
| # define avail_ASR17(C) false |
| # define avail_CASA(C) true |
| # define avail_DIV(C) true |
| # define avail_MUL(C) true |
| # define avail_POWERDOWN(C) false |
| # define avail_64(C) true |
| # define avail_FMAF(C) ((C)->def->features & CPU_FEATURE_FMAF) |
| # define avail_GL(C) ((C)->def->features & CPU_FEATURE_GL) |
| # define avail_HYPV(C) ((C)->def->features & CPU_FEATURE_HYPV) |
| # define avail_IMA(C) ((C)->def->features & CPU_FEATURE_IMA) |
| # define avail_VIS1(C) ((C)->def->features & CPU_FEATURE_VIS1) |
| # define avail_VIS2(C) ((C)->def->features & CPU_FEATURE_VIS2) |
| # define avail_VIS3(C) ((C)->def->features & CPU_FEATURE_VIS3) |
| # define avail_VIS3B(C) avail_VIS3(C) |
| # define avail_VIS4(C) ((C)->def->features & CPU_FEATURE_VIS4) |
| #else |
| # define avail_32(C) true |
| # define avail_ASR17(C) ((C)->def->features & CPU_FEATURE_ASR17) |
| # define avail_CASA(C) ((C)->def->features & CPU_FEATURE_CASA) |
| # define avail_DIV(C) ((C)->def->features & CPU_FEATURE_DIV) |
| # define avail_MUL(C) ((C)->def->features & CPU_FEATURE_MUL) |
| # define avail_POWERDOWN(C) ((C)->def->features & CPU_FEATURE_POWERDOWN) |
| # define avail_64(C) false |
| # define avail_FMAF(C) false |
| # define avail_GL(C) false |
| # define avail_HYPV(C) false |
| # define avail_IMA(C) false |
| # define avail_VIS1(C) false |
| # define avail_VIS2(C) false |
| # define avail_VIS3(C) false |
| # define avail_VIS3B(C) false |
| # define avail_VIS4(C) false |
| #endif |
| |
| /* Default case for non jump instructions. */ |
| static bool advance_pc(DisasContext *dc) |
| { |
| TCGLabel *l1; |
| |
| finishing_insn(dc); |
| |
| if (dc->npc & 3) { |
| switch (dc->npc) { |
| case DYNAMIC_PC: |
| case DYNAMIC_PC_LOOKUP: |
| dc->pc = dc->npc; |
| tcg_gen_mov_tl(cpu_pc, cpu_npc); |
| tcg_gen_addi_tl(cpu_npc, cpu_npc, 4); |
| break; |
| |
| case JUMP_PC: |
| /* we can do a static jump */ |
| l1 = gen_new_label(); |
| tcg_gen_brcondi_tl(dc->jump.cond, dc->jump.c1, dc->jump.c2, l1); |
| |
| /* jump not taken */ |
| gen_goto_tb(dc, 1, dc->jump_pc[1], dc->jump_pc[1] + 4); |
| |
| /* jump taken */ |
| gen_set_label(l1); |
| gen_goto_tb(dc, 0, dc->jump_pc[0], dc->jump_pc[0] + 4); |
| |
| dc->base.is_jmp = DISAS_NORETURN; |
| break; |
| |
| default: |
| g_assert_not_reached(); |
| } |
| } else { |
| dc->pc = dc->npc; |
| dc->npc = dc->npc + 4; |
| } |
| return true; |
| } |
| |
| /* |
| * Major opcodes 00 and 01 -- branches, call, and sethi |
| */ |
| |
| static bool advance_jump_cond(DisasContext *dc, DisasCompare *cmp, |
| bool annul, int disp) |
| { |
| target_ulong dest = address_mask_i(dc, dc->pc + disp * 4); |
| target_ulong npc; |
| |
| finishing_insn(dc); |
| |
| if (cmp->cond == TCG_COND_ALWAYS) { |
| if (annul) { |
| dc->pc = dest; |
| dc->npc = dest + 4; |
| } else { |
| gen_mov_pc_npc(dc); |
| dc->npc = dest; |
| } |
| return true; |
| } |
| |
| if (cmp->cond == TCG_COND_NEVER) { |
| npc = dc->npc; |
| if (npc & 3) { |
| gen_mov_pc_npc(dc); |
| if (annul) { |
| tcg_gen_addi_tl(cpu_pc, cpu_pc, 4); |
| } |
| tcg_gen_addi_tl(cpu_npc, cpu_pc, 4); |
| } else { |
| dc->pc = npc + (annul ? 4 : 0); |
| dc->npc = dc->pc + 4; |
| } |
| return true; |
| } |
| |
| flush_cond(dc); |
| npc = dc->npc; |
| |
| if (annul) { |
| TCGLabel *l1 = gen_new_label(); |
| |
| tcg_gen_brcondi_tl(tcg_invert_cond(cmp->cond), cmp->c1, cmp->c2, l1); |
| gen_goto_tb(dc, 0, npc, dest); |
| gen_set_label(l1); |
| gen_goto_tb(dc, 1, npc + 4, npc + 8); |
| |
| dc->base.is_jmp = DISAS_NORETURN; |
| } else { |
| if (npc & 3) { |
| switch (npc) { |
| case DYNAMIC_PC: |
| case DYNAMIC_PC_LOOKUP: |
| tcg_gen_mov_tl(cpu_pc, cpu_npc); |
| tcg_gen_addi_tl(cpu_npc, cpu_npc, 4); |
| tcg_gen_movcond_tl(cmp->cond, cpu_npc, |
| cmp->c1, tcg_constant_tl(cmp->c2), |
| tcg_constant_tl(dest), cpu_npc); |
| dc->pc = npc; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } else { |
| dc->pc = npc; |
| dc->npc = JUMP_PC; |
| dc->jump = *cmp; |
| dc->jump_pc[0] = dest; |
| dc->jump_pc[1] = npc + 4; |
| |
| /* The condition for cpu_cond is always NE -- normalize. */ |
| if (cmp->cond == TCG_COND_NE) { |
| tcg_gen_xori_tl(cpu_cond, cmp->c1, cmp->c2); |
| } else { |
| tcg_gen_setcondi_tl(cmp->cond, cpu_cond, cmp->c1, cmp->c2); |
| } |
| dc->cpu_cond_live = true; |
| } |
| } |
| return true; |
| } |
| |
| static bool raise_priv(DisasContext *dc) |
| { |
| gen_exception(dc, TT_PRIV_INSN); |
| return true; |
| } |
| |
| static bool raise_unimpfpop(DisasContext *dc) |
| { |
| gen_op_fpexception_im(dc, FSR_FTT_UNIMPFPOP); |
| return true; |
| } |
| |
| static bool gen_trap_float128(DisasContext *dc) |
| { |
| if (dc->def->features & CPU_FEATURE_FLOAT128) { |
| return false; |
| } |
| return raise_unimpfpop(dc); |
| } |
| |
| static bool do_bpcc(DisasContext *dc, arg_bcc *a) |
| { |
| DisasCompare cmp; |
| |
| gen_compare(&cmp, a->cc, a->cond, dc); |
| return advance_jump_cond(dc, &cmp, a->a, a->i); |
| } |
| |
| TRANS(Bicc, ALL, do_bpcc, a) |
| TRANS(BPcc, 64, do_bpcc, a) |
| |
| static bool do_fbpfcc(DisasContext *dc, arg_bcc *a) |
| { |
| DisasCompare cmp; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| gen_fcompare(&cmp, a->cc, a->cond); |
| return advance_jump_cond(dc, &cmp, a->a, a->i); |
| } |
| |
| TRANS(FBPfcc, 64, do_fbpfcc, a) |
| TRANS(FBfcc, ALL, do_fbpfcc, a) |
| |
| static bool trans_BPr(DisasContext *dc, arg_BPr *a) |
| { |
| DisasCompare cmp; |
| |
| if (!avail_64(dc)) { |
| return false; |
| } |
| if (!gen_compare_reg(&cmp, a->cond, gen_load_gpr(dc, a->rs1))) { |
| return false; |
| } |
| return advance_jump_cond(dc, &cmp, a->a, a->i); |
| } |
| |
| static bool trans_CALL(DisasContext *dc, arg_CALL *a) |
| { |
| target_long target = address_mask_i(dc, dc->pc + a->i * 4); |
| |
| gen_store_gpr(dc, 15, tcg_constant_tl(dc->pc)); |
| gen_mov_pc_npc(dc); |
| dc->npc = target; |
| return true; |
| } |
| |
| static bool trans_NCP(DisasContext *dc, arg_NCP *a) |
| { |
| /* |
| * For sparc32, always generate the no-coprocessor exception. |
| * For sparc64, always generate illegal instruction. |
| */ |
| #ifdef TARGET_SPARC64 |
| return false; |
| #else |
| gen_exception(dc, TT_NCP_INSN); |
| return true; |
| #endif |
| } |
| |
| static bool trans_SETHI(DisasContext *dc, arg_SETHI *a) |
| { |
| /* Special-case %g0 because that's the canonical nop. */ |
| if (a->rd) { |
| gen_store_gpr(dc, a->rd, tcg_constant_tl((uint32_t)a->i << 10)); |
| } |
| return advance_pc(dc); |
| } |
| |
| /* |
| * Major Opcode 10 -- integer, floating-point, vis, and system insns. |
| */ |
| |
| static bool do_tcc(DisasContext *dc, int cond, int cc, |
| int rs1, bool imm, int rs2_or_imm) |
| { |
| int mask = ((dc->def->features & CPU_FEATURE_HYPV) && supervisor(dc) |
| ? UA2005_HTRAP_MASK : V8_TRAP_MASK); |
| DisasCompare cmp; |
| TCGLabel *lab; |
| TCGv_i32 trap; |
| |
| /* Trap never. */ |
| if (cond == 0) { |
| return advance_pc(dc); |
| } |
| |
| /* |
| * Immediate traps are the most common case. Since this value is |
| * live across the branch, it really pays to evaluate the constant. |
| */ |
| if (rs1 == 0 && (imm || rs2_or_imm == 0)) { |
| trap = tcg_constant_i32((rs2_or_imm & mask) + TT_TRAP); |
| } else { |
| trap = tcg_temp_new_i32(); |
| tcg_gen_trunc_tl_i32(trap, gen_load_gpr(dc, rs1)); |
| if (imm) { |
| tcg_gen_addi_i32(trap, trap, rs2_or_imm); |
| } else { |
| TCGv_i32 t2 = tcg_temp_new_i32(); |
| tcg_gen_trunc_tl_i32(t2, gen_load_gpr(dc, rs2_or_imm)); |
| tcg_gen_add_i32(trap, trap, t2); |
| } |
| tcg_gen_andi_i32(trap, trap, mask); |
| tcg_gen_addi_i32(trap, trap, TT_TRAP); |
| } |
| |
| finishing_insn(dc); |
| |
| /* Trap always. */ |
| if (cond == 8) { |
| save_state(dc); |
| gen_helper_raise_exception(tcg_env, trap); |
| dc->base.is_jmp = DISAS_NORETURN; |
| return true; |
| } |
| |
| /* Conditional trap. */ |
| flush_cond(dc); |
| lab = delay_exceptionv(dc, trap); |
| gen_compare(&cmp, cc, cond, dc); |
| tcg_gen_brcondi_tl(cmp.cond, cmp.c1, cmp.c2, lab); |
| |
| return advance_pc(dc); |
| } |
| |
| static bool trans_Tcc_r(DisasContext *dc, arg_Tcc_r *a) |
| { |
| if (avail_32(dc) && a->cc) { |
| return false; |
| } |
| return do_tcc(dc, a->cond, a->cc, a->rs1, false, a->rs2); |
| } |
| |
| static bool trans_Tcc_i_v7(DisasContext *dc, arg_Tcc_i_v7 *a) |
| { |
| if (avail_64(dc)) { |
| return false; |
| } |
| return do_tcc(dc, a->cond, 0, a->rs1, true, a->i); |
| } |
| |
| static bool trans_Tcc_i_v9(DisasContext *dc, arg_Tcc_i_v9 *a) |
| { |
| if (avail_32(dc)) { |
| return false; |
| } |
| return do_tcc(dc, a->cond, a->cc, a->rs1, true, a->i); |
| } |
| |
| static bool trans_STBAR(DisasContext *dc, arg_STBAR *a) |
| { |
| tcg_gen_mb(TCG_MO_ST_ST | TCG_BAR_SC); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_MEMBAR(DisasContext *dc, arg_MEMBAR *a) |
| { |
| if (avail_32(dc)) { |
| return false; |
| } |
| if (a->mmask) { |
| /* Note TCG_MO_* was modeled on sparc64, so mmask matches. */ |
| tcg_gen_mb(a->mmask | TCG_BAR_SC); |
| } |
| if (a->cmask) { |
| /* For #Sync, etc, end the TB to recognize interrupts. */ |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| return advance_pc(dc); |
| } |
| |
| static bool do_rd_special(DisasContext *dc, bool priv, int rd, |
| TCGv (*func)(DisasContext *, TCGv)) |
| { |
| if (!priv) { |
| return raise_priv(dc); |
| } |
| gen_store_gpr(dc, rd, func(dc, gen_dest_gpr(dc, rd))); |
| return advance_pc(dc); |
| } |
| |
| static TCGv do_rdy(DisasContext *dc, TCGv dst) |
| { |
| return cpu_y; |
| } |
| |
| static bool trans_RDY(DisasContext *dc, arg_RDY *a) |
| { |
| /* |
| * TODO: Need a feature bit for sparcv8. In the meantime, treat all |
| * 32-bit cpus like sparcv7, which ignores the rs1 field. |
| * This matches after all other ASR, so Leon3 Asr17 is handled first. |
| */ |
| if (avail_64(dc) && a->rs1 != 0) { |
| return false; |
| } |
| return do_rd_special(dc, true, a->rd, do_rdy); |
| } |
| |
| static TCGv do_rd_leon3_config(DisasContext *dc, TCGv dst) |
| { |
| gen_helper_rdasr17(dst, tcg_env); |
| return dst; |
| } |
| |
| TRANS(RDASR17, ASR17, do_rd_special, true, a->rd, do_rd_leon3_config) |
| |
| static TCGv do_rdccr(DisasContext *dc, TCGv dst) |
| { |
| gen_helper_rdccr(dst, tcg_env); |
| return dst; |
| } |
| |
| TRANS(RDCCR, 64, do_rd_special, true, a->rd, do_rdccr) |
| |
| static TCGv do_rdasi(DisasContext *dc, TCGv dst) |
| { |
| #ifdef TARGET_SPARC64 |
| return tcg_constant_tl(dc->asi); |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(RDASI, 64, do_rd_special, true, a->rd, do_rdasi) |
| |
| static TCGv do_rdtick(DisasContext *dc, TCGv dst) |
| { |
| TCGv_ptr r_tickptr = tcg_temp_new_ptr(); |
| |
| tcg_gen_ld_ptr(r_tickptr, tcg_env, env64_field_offsetof(tick)); |
| if (translator_io_start(&dc->base)) { |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| gen_helper_tick_get_count(dst, tcg_env, r_tickptr, |
| tcg_constant_i32(dc->mem_idx)); |
| return dst; |
| } |
| |
| /* TODO: non-priv access only allowed when enabled. */ |
| TRANS(RDTICK, 64, do_rd_special, true, a->rd, do_rdtick) |
| |
| static TCGv do_rdpc(DisasContext *dc, TCGv dst) |
| { |
| return tcg_constant_tl(address_mask_i(dc, dc->pc)); |
| } |
| |
| TRANS(RDPC, 64, do_rd_special, true, a->rd, do_rdpc) |
| |
| static TCGv do_rdfprs(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ext_i32_tl(dst, cpu_fprs); |
| return dst; |
| } |
| |
| TRANS(RDFPRS, 64, do_rd_special, true, a->rd, do_rdfprs) |
| |
| static TCGv do_rdgsr(DisasContext *dc, TCGv dst) |
| { |
| gen_trap_ifnofpu(dc); |
| return cpu_gsr; |
| } |
| |
| TRANS(RDGSR, 64, do_rd_special, true, a->rd, do_rdgsr) |
| |
| static TCGv do_rdsoftint(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(softint)); |
| return dst; |
| } |
| |
| TRANS(RDSOFTINT, 64, do_rd_special, supervisor(dc), a->rd, do_rdsoftint) |
| |
| static TCGv do_rdtick_cmpr(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(tick_cmpr)); |
| return dst; |
| } |
| |
| /* TODO: non-priv access only allowed when enabled. */ |
| TRANS(RDTICK_CMPR, 64, do_rd_special, true, a->rd, do_rdtick_cmpr) |
| |
| static TCGv do_rdstick(DisasContext *dc, TCGv dst) |
| { |
| TCGv_ptr r_tickptr = tcg_temp_new_ptr(); |
| |
| tcg_gen_ld_ptr(r_tickptr, tcg_env, env64_field_offsetof(stick)); |
| if (translator_io_start(&dc->base)) { |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| gen_helper_tick_get_count(dst, tcg_env, r_tickptr, |
| tcg_constant_i32(dc->mem_idx)); |
| return dst; |
| } |
| |
| /* TODO: non-priv access only allowed when enabled. */ |
| TRANS(RDSTICK, 64, do_rd_special, true, a->rd, do_rdstick) |
| |
| static TCGv do_rdstick_cmpr(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(stick_cmpr)); |
| return dst; |
| } |
| |
| /* TODO: supervisor access only allowed when enabled by hypervisor. */ |
| TRANS(RDSTICK_CMPR, 64, do_rd_special, supervisor(dc), a->rd, do_rdstick_cmpr) |
| |
| /* |
| * UltraSPARC-T1 Strand status. |
| * HYPV check maybe not enough, UA2005 & UA2007 describe |
| * this ASR as impl. dep |
| */ |
| static TCGv do_rdstrand_status(DisasContext *dc, TCGv dst) |
| { |
| return tcg_constant_tl(1); |
| } |
| |
| TRANS(RDSTRAND_STATUS, HYPV, do_rd_special, true, a->rd, do_rdstrand_status) |
| |
| static TCGv do_rdpsr(DisasContext *dc, TCGv dst) |
| { |
| gen_helper_rdpsr(dst, tcg_env); |
| return dst; |
| } |
| |
| TRANS(RDPSR, 32, do_rd_special, supervisor(dc), a->rd, do_rdpsr) |
| |
| static TCGv do_rdhpstate(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(hpstate)); |
| return dst; |
| } |
| |
| TRANS(RDHPR_hpstate, HYPV, do_rd_special, hypervisor(dc), a->rd, do_rdhpstate) |
| |
| static TCGv do_rdhtstate(DisasContext *dc, TCGv dst) |
| { |
| TCGv_i32 tl = tcg_temp_new_i32(); |
| TCGv_ptr tp = tcg_temp_new_ptr(); |
| |
| tcg_gen_ld_i32(tl, tcg_env, env64_field_offsetof(tl)); |
| tcg_gen_andi_i32(tl, tl, MAXTL_MASK); |
| tcg_gen_shli_i32(tl, tl, 3); |
| tcg_gen_ext_i32_ptr(tp, tl); |
| tcg_gen_add_ptr(tp, tp, tcg_env); |
| |
| tcg_gen_ld_tl(dst, tp, env64_field_offsetof(htstate)); |
| return dst; |
| } |
| |
| TRANS(RDHPR_htstate, HYPV, do_rd_special, hypervisor(dc), a->rd, do_rdhtstate) |
| |
| static TCGv do_rdhintp(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(hintp)); |
| return dst; |
| } |
| |
| TRANS(RDHPR_hintp, HYPV, do_rd_special, hypervisor(dc), a->rd, do_rdhintp) |
| |
| static TCGv do_rdhtba(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(htba)); |
| return dst; |
| } |
| |
| TRANS(RDHPR_htba, HYPV, do_rd_special, hypervisor(dc), a->rd, do_rdhtba) |
| |
| static TCGv do_rdhver(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(hver)); |
| return dst; |
| } |
| |
| TRANS(RDHPR_hver, HYPV, do_rd_special, hypervisor(dc), a->rd, do_rdhver) |
| |
| static TCGv do_rdhstick_cmpr(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(hstick_cmpr)); |
| return dst; |
| } |
| |
| TRANS(RDHPR_hstick_cmpr, HYPV, do_rd_special, hypervisor(dc), a->rd, |
| do_rdhstick_cmpr) |
| |
| static TCGv do_rdwim(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env32_field_offsetof(wim)); |
| return dst; |
| } |
| |
| TRANS(RDWIM, 32, do_rd_special, supervisor(dc), a->rd, do_rdwim) |
| |
| static TCGv do_rdtpc(DisasContext *dc, TCGv dst) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tsptr = tcg_temp_new_ptr(); |
| |
| gen_load_trap_state_at_tl(r_tsptr); |
| tcg_gen_ld_tl(dst, r_tsptr, offsetof(trap_state, tpc)); |
| return dst; |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(RDPR_tpc, 64, do_rd_special, supervisor(dc), a->rd, do_rdtpc) |
| |
| static TCGv do_rdtnpc(DisasContext *dc, TCGv dst) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tsptr = tcg_temp_new_ptr(); |
| |
| gen_load_trap_state_at_tl(r_tsptr); |
| tcg_gen_ld_tl(dst, r_tsptr, offsetof(trap_state, tnpc)); |
| return dst; |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(RDPR_tnpc, 64, do_rd_special, supervisor(dc), a->rd, do_rdtnpc) |
| |
| static TCGv do_rdtstate(DisasContext *dc, TCGv dst) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tsptr = tcg_temp_new_ptr(); |
| |
| gen_load_trap_state_at_tl(r_tsptr); |
| tcg_gen_ld_tl(dst, r_tsptr, offsetof(trap_state, tstate)); |
| return dst; |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(RDPR_tstate, 64, do_rd_special, supervisor(dc), a->rd, do_rdtstate) |
| |
| static TCGv do_rdtt(DisasContext *dc, TCGv dst) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tsptr = tcg_temp_new_ptr(); |
| |
| gen_load_trap_state_at_tl(r_tsptr); |
| tcg_gen_ld32s_tl(dst, r_tsptr, offsetof(trap_state, tt)); |
| return dst; |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(RDPR_tt, 64, do_rd_special, supervisor(dc), a->rd, do_rdtt) |
| TRANS(RDPR_tick, 64, do_rd_special, supervisor(dc), a->rd, do_rdtick) |
| |
| static TCGv do_rdtba(DisasContext *dc, TCGv dst) |
| { |
| return cpu_tbr; |
| } |
| |
| TRANS(RDTBR, 32, do_rd_special, supervisor(dc), a->rd, do_rdtba) |
| TRANS(RDPR_tba, 64, do_rd_special, supervisor(dc), a->rd, do_rdtba) |
| |
| static TCGv do_rdpstate(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(pstate)); |
| return dst; |
| } |
| |
| TRANS(RDPR_pstate, 64, do_rd_special, supervisor(dc), a->rd, do_rdpstate) |
| |
| static TCGv do_rdtl(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(tl)); |
| return dst; |
| } |
| |
| TRANS(RDPR_tl, 64, do_rd_special, supervisor(dc), a->rd, do_rdtl) |
| |
| static TCGv do_rdpil(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env_field_offsetof(psrpil)); |
| return dst; |
| } |
| |
| TRANS(RDPR_pil, 64, do_rd_special, supervisor(dc), a->rd, do_rdpil) |
| |
| static TCGv do_rdcwp(DisasContext *dc, TCGv dst) |
| { |
| gen_helper_rdcwp(dst, tcg_env); |
| return dst; |
| } |
| |
| TRANS(RDPR_cwp, 64, do_rd_special, supervisor(dc), a->rd, do_rdcwp) |
| |
| static TCGv do_rdcansave(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(cansave)); |
| return dst; |
| } |
| |
| TRANS(RDPR_cansave, 64, do_rd_special, supervisor(dc), a->rd, do_rdcansave) |
| |
| static TCGv do_rdcanrestore(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(canrestore)); |
| return dst; |
| } |
| |
| TRANS(RDPR_canrestore, 64, do_rd_special, supervisor(dc), a->rd, |
| do_rdcanrestore) |
| |
| static TCGv do_rdcleanwin(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(cleanwin)); |
| return dst; |
| } |
| |
| TRANS(RDPR_cleanwin, 64, do_rd_special, supervisor(dc), a->rd, do_rdcleanwin) |
| |
| static TCGv do_rdotherwin(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(otherwin)); |
| return dst; |
| } |
| |
| TRANS(RDPR_otherwin, 64, do_rd_special, supervisor(dc), a->rd, do_rdotherwin) |
| |
| static TCGv do_rdwstate(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(wstate)); |
| return dst; |
| } |
| |
| TRANS(RDPR_wstate, 64, do_rd_special, supervisor(dc), a->rd, do_rdwstate) |
| |
| static TCGv do_rdgl(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld32s_tl(dst, tcg_env, env64_field_offsetof(gl)); |
| return dst; |
| } |
| |
| TRANS(RDPR_gl, GL, do_rd_special, supervisor(dc), a->rd, do_rdgl) |
| |
| /* UA2005 strand status */ |
| static TCGv do_rdssr(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(ssr)); |
| return dst; |
| } |
| |
| TRANS(RDPR_strand_status, HYPV, do_rd_special, hypervisor(dc), a->rd, do_rdssr) |
| |
| static TCGv do_rdver(DisasContext *dc, TCGv dst) |
| { |
| tcg_gen_ld_tl(dst, tcg_env, env64_field_offsetof(version)); |
| return dst; |
| } |
| |
| TRANS(RDPR_ver, 64, do_rd_special, supervisor(dc), a->rd, do_rdver) |
| |
| static bool trans_FLUSHW(DisasContext *dc, arg_FLUSHW *a) |
| { |
| if (avail_64(dc)) { |
| gen_helper_flushw(tcg_env); |
| return advance_pc(dc); |
| } |
| return false; |
| } |
| |
| static bool do_wr_special(DisasContext *dc, arg_r_r_ri *a, bool priv, |
| void (*func)(DisasContext *, TCGv)) |
| { |
| TCGv src; |
| |
| /* For simplicity, we under-decoded the rs2 form. */ |
| if (!a->imm && (a->rs2_or_imm & ~0x1f)) { |
| return false; |
| } |
| if (!priv) { |
| return raise_priv(dc); |
| } |
| |
| if (a->rs1 == 0 && (a->imm || a->rs2_or_imm == 0)) { |
| src = tcg_constant_tl(a->rs2_or_imm); |
| } else { |
| TCGv src1 = gen_load_gpr(dc, a->rs1); |
| if (a->rs2_or_imm == 0) { |
| src = src1; |
| } else { |
| src = tcg_temp_new(); |
| if (a->imm) { |
| tcg_gen_xori_tl(src, src1, a->rs2_or_imm); |
| } else { |
| tcg_gen_xor_tl(src, src1, gen_load_gpr(dc, a->rs2_or_imm)); |
| } |
| } |
| } |
| func(dc, src); |
| return advance_pc(dc); |
| } |
| |
| static void do_wry(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_ext32u_tl(cpu_y, src); |
| } |
| |
| TRANS(WRY, ALL, do_wr_special, a, true, do_wry) |
| |
| static void do_wrccr(DisasContext *dc, TCGv src) |
| { |
| gen_helper_wrccr(tcg_env, src); |
| } |
| |
| TRANS(WRCCR, 64, do_wr_special, a, true, do_wrccr) |
| |
| static void do_wrasi(DisasContext *dc, TCGv src) |
| { |
| TCGv tmp = tcg_temp_new(); |
| |
| tcg_gen_ext8u_tl(tmp, src); |
| tcg_gen_st32_tl(tmp, tcg_env, env64_field_offsetof(asi)); |
| /* End TB to notice changed ASI. */ |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| |
| TRANS(WRASI, 64, do_wr_special, a, true, do_wrasi) |
| |
| static void do_wrfprs(DisasContext *dc, TCGv src) |
| { |
| #ifdef TARGET_SPARC64 |
| tcg_gen_trunc_tl_i32(cpu_fprs, src); |
| dc->fprs_dirty = 0; |
| dc->base.is_jmp = DISAS_EXIT; |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(WRFPRS, 64, do_wr_special, a, true, do_wrfprs) |
| |
| static void do_wrgsr(DisasContext *dc, TCGv src) |
| { |
| gen_trap_ifnofpu(dc); |
| tcg_gen_mov_tl(cpu_gsr, src); |
| } |
| |
| TRANS(WRGSR, 64, do_wr_special, a, true, do_wrgsr) |
| |
| static void do_wrsoftint_set(DisasContext *dc, TCGv src) |
| { |
| gen_helper_set_softint(tcg_env, src); |
| } |
| |
| TRANS(WRSOFTINT_SET, 64, do_wr_special, a, supervisor(dc), do_wrsoftint_set) |
| |
| static void do_wrsoftint_clr(DisasContext *dc, TCGv src) |
| { |
| gen_helper_clear_softint(tcg_env, src); |
| } |
| |
| TRANS(WRSOFTINT_CLR, 64, do_wr_special, a, supervisor(dc), do_wrsoftint_clr) |
| |
| static void do_wrsoftint(DisasContext *dc, TCGv src) |
| { |
| gen_helper_write_softint(tcg_env, src); |
| } |
| |
| TRANS(WRSOFTINT, 64, do_wr_special, a, supervisor(dc), do_wrsoftint) |
| |
| static void do_wrtick_cmpr(DisasContext *dc, TCGv src) |
| { |
| TCGv_ptr r_tickptr = tcg_temp_new_ptr(); |
| |
| tcg_gen_st_tl(src, tcg_env, env64_field_offsetof(tick_cmpr)); |
| tcg_gen_ld_ptr(r_tickptr, tcg_env, env64_field_offsetof(tick)); |
| translator_io_start(&dc->base); |
| gen_helper_tick_set_limit(r_tickptr, src); |
| /* End TB to handle timer interrupt */ |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| |
| TRANS(WRTICK_CMPR, 64, do_wr_special, a, supervisor(dc), do_wrtick_cmpr) |
| |
| static void do_wrstick(DisasContext *dc, TCGv src) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tickptr = tcg_temp_new_ptr(); |
| |
| tcg_gen_ld_ptr(r_tickptr, tcg_env, offsetof(CPUSPARCState, stick)); |
| translator_io_start(&dc->base); |
| gen_helper_tick_set_count(r_tickptr, src); |
| /* End TB to handle timer interrupt */ |
| dc->base.is_jmp = DISAS_EXIT; |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(WRSTICK, 64, do_wr_special, a, supervisor(dc), do_wrstick) |
| |
| static void do_wrstick_cmpr(DisasContext *dc, TCGv src) |
| { |
| TCGv_ptr r_tickptr = tcg_temp_new_ptr(); |
| |
| tcg_gen_st_tl(src, tcg_env, env64_field_offsetof(stick_cmpr)); |
| tcg_gen_ld_ptr(r_tickptr, tcg_env, env64_field_offsetof(stick)); |
| translator_io_start(&dc->base); |
| gen_helper_tick_set_limit(r_tickptr, src); |
| /* End TB to handle timer interrupt */ |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| |
| TRANS(WRSTICK_CMPR, 64, do_wr_special, a, supervisor(dc), do_wrstick_cmpr) |
| |
| static void do_wrpowerdown(DisasContext *dc, TCGv src) |
| { |
| finishing_insn(dc); |
| save_state(dc); |
| gen_helper_power_down(tcg_env); |
| } |
| |
| TRANS(WRPOWERDOWN, POWERDOWN, do_wr_special, a, supervisor(dc), do_wrpowerdown) |
| |
| static void do_wrmwait(DisasContext *dc, TCGv src) |
| { |
| /* |
| * TODO: This is a stub version of mwait, which merely recognizes |
| * interrupts immediately and does not wait. |
| */ |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| |
| TRANS(WRMWAIT, VIS4, do_wr_special, a, true, do_wrmwait) |
| |
| static void do_wrpsr(DisasContext *dc, TCGv src) |
| { |
| gen_helper_wrpsr(tcg_env, src); |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| |
| TRANS(WRPSR, 32, do_wr_special, a, supervisor(dc), do_wrpsr) |
| |
| static void do_wrwim(DisasContext *dc, TCGv src) |
| { |
| target_ulong mask = MAKE_64BIT_MASK(0, dc->def->nwindows); |
| TCGv tmp = tcg_temp_new(); |
| |
| tcg_gen_andi_tl(tmp, src, mask); |
| tcg_gen_st_tl(tmp, tcg_env, env32_field_offsetof(wim)); |
| } |
| |
| TRANS(WRWIM, 32, do_wr_special, a, supervisor(dc), do_wrwim) |
| |
| static void do_wrtpc(DisasContext *dc, TCGv src) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tsptr = tcg_temp_new_ptr(); |
| |
| gen_load_trap_state_at_tl(r_tsptr); |
| tcg_gen_st_tl(src, r_tsptr, offsetof(trap_state, tpc)); |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(WRPR_tpc, 64, do_wr_special, a, supervisor(dc), do_wrtpc) |
| |
| static void do_wrtnpc(DisasContext *dc, TCGv src) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tsptr = tcg_temp_new_ptr(); |
| |
| gen_load_trap_state_at_tl(r_tsptr); |
| tcg_gen_st_tl(src, r_tsptr, offsetof(trap_state, tnpc)); |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(WRPR_tnpc, 64, do_wr_special, a, supervisor(dc), do_wrtnpc) |
| |
| static void do_wrtstate(DisasContext *dc, TCGv src) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tsptr = tcg_temp_new_ptr(); |
| |
| gen_load_trap_state_at_tl(r_tsptr); |
| tcg_gen_st_tl(src, r_tsptr, offsetof(trap_state, tstate)); |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(WRPR_tstate, 64, do_wr_special, a, supervisor(dc), do_wrtstate) |
| |
| static void do_wrtt(DisasContext *dc, TCGv src) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv_ptr r_tsptr = tcg_temp_new_ptr(); |
| |
| gen_load_trap_state_at_tl(r_tsptr); |
| tcg_gen_st32_tl(src, r_tsptr, offsetof(trap_state, tt)); |
| #else |
| qemu_build_not_reached(); |
| #endif |
| } |
| |
| TRANS(WRPR_tt, 64, do_wr_special, a, supervisor(dc), do_wrtt) |
| |
| static void do_wrtick(DisasContext *dc, TCGv src) |
| { |
| TCGv_ptr r_tickptr = tcg_temp_new_ptr(); |
| |
| tcg_gen_ld_ptr(r_tickptr, tcg_env, env64_field_offsetof(tick)); |
| translator_io_start(&dc->base); |
| gen_helper_tick_set_count(r_tickptr, src); |
| /* End TB to handle timer interrupt */ |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| |
| TRANS(WRPR_tick, 64, do_wr_special, a, supervisor(dc), do_wrtick) |
| |
| static void do_wrtba(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_mov_tl(cpu_tbr, src); |
| } |
| |
| TRANS(WRPR_tba, 64, do_wr_special, a, supervisor(dc), do_wrtba) |
| |
| static void do_wrpstate(DisasContext *dc, TCGv src) |
| { |
| save_state(dc); |
| if (translator_io_start(&dc->base)) { |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| gen_helper_wrpstate(tcg_env, src); |
| dc->npc = DYNAMIC_PC; |
| } |
| |
| TRANS(WRPR_pstate, 64, do_wr_special, a, supervisor(dc), do_wrpstate) |
| |
| static void do_wrtl(DisasContext *dc, TCGv src) |
| { |
| save_state(dc); |
| tcg_gen_st32_tl(src, tcg_env, env64_field_offsetof(tl)); |
| dc->npc = DYNAMIC_PC; |
| } |
| |
| TRANS(WRPR_tl, 64, do_wr_special, a, supervisor(dc), do_wrtl) |
| |
| static void do_wrpil(DisasContext *dc, TCGv src) |
| { |
| if (translator_io_start(&dc->base)) { |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| gen_helper_wrpil(tcg_env, src); |
| } |
| |
| TRANS(WRPR_pil, 64, do_wr_special, a, supervisor(dc), do_wrpil) |
| |
| static void do_wrcwp(DisasContext *dc, TCGv src) |
| { |
| gen_helper_wrcwp(tcg_env, src); |
| } |
| |
| TRANS(WRPR_cwp, 64, do_wr_special, a, supervisor(dc), do_wrcwp) |
| |
| static void do_wrcansave(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st32_tl(src, tcg_env, env64_field_offsetof(cansave)); |
| } |
| |
| TRANS(WRPR_cansave, 64, do_wr_special, a, supervisor(dc), do_wrcansave) |
| |
| static void do_wrcanrestore(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st32_tl(src, tcg_env, env64_field_offsetof(canrestore)); |
| } |
| |
| TRANS(WRPR_canrestore, 64, do_wr_special, a, supervisor(dc), do_wrcanrestore) |
| |
| static void do_wrcleanwin(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st32_tl(src, tcg_env, env64_field_offsetof(cleanwin)); |
| } |
| |
| TRANS(WRPR_cleanwin, 64, do_wr_special, a, supervisor(dc), do_wrcleanwin) |
| |
| static void do_wrotherwin(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st32_tl(src, tcg_env, env64_field_offsetof(otherwin)); |
| } |
| |
| TRANS(WRPR_otherwin, 64, do_wr_special, a, supervisor(dc), do_wrotherwin) |
| |
| static void do_wrwstate(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st32_tl(src, tcg_env, env64_field_offsetof(wstate)); |
| } |
| |
| TRANS(WRPR_wstate, 64, do_wr_special, a, supervisor(dc), do_wrwstate) |
| |
| static void do_wrgl(DisasContext *dc, TCGv src) |
| { |
| gen_helper_wrgl(tcg_env, src); |
| } |
| |
| TRANS(WRPR_gl, GL, do_wr_special, a, supervisor(dc), do_wrgl) |
| |
| /* UA2005 strand status */ |
| static void do_wrssr(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st_tl(src, tcg_env, env64_field_offsetof(ssr)); |
| } |
| |
| TRANS(WRPR_strand_status, HYPV, do_wr_special, a, hypervisor(dc), do_wrssr) |
| |
| TRANS(WRTBR, 32, do_wr_special, a, supervisor(dc), do_wrtba) |
| |
| static void do_wrhpstate(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st_tl(src, tcg_env, env64_field_offsetof(hpstate)); |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| |
| TRANS(WRHPR_hpstate, HYPV, do_wr_special, a, hypervisor(dc), do_wrhpstate) |
| |
| static void do_wrhtstate(DisasContext *dc, TCGv src) |
| { |
| TCGv_i32 tl = tcg_temp_new_i32(); |
| TCGv_ptr tp = tcg_temp_new_ptr(); |
| |
| tcg_gen_ld_i32(tl, tcg_env, env64_field_offsetof(tl)); |
| tcg_gen_andi_i32(tl, tl, MAXTL_MASK); |
| tcg_gen_shli_i32(tl, tl, 3); |
| tcg_gen_ext_i32_ptr(tp, tl); |
| tcg_gen_add_ptr(tp, tp, tcg_env); |
| |
| tcg_gen_st_tl(src, tp, env64_field_offsetof(htstate)); |
| } |
| |
| TRANS(WRHPR_htstate, HYPV, do_wr_special, a, hypervisor(dc), do_wrhtstate) |
| |
| static void do_wrhintp(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st_tl(src, tcg_env, env64_field_offsetof(hintp)); |
| } |
| |
| TRANS(WRHPR_hintp, HYPV, do_wr_special, a, hypervisor(dc), do_wrhintp) |
| |
| static void do_wrhtba(DisasContext *dc, TCGv src) |
| { |
| tcg_gen_st_tl(src, tcg_env, env64_field_offsetof(htba)); |
| } |
| |
| TRANS(WRHPR_htba, HYPV, do_wr_special, a, hypervisor(dc), do_wrhtba) |
| |
| static void do_wrhstick_cmpr(DisasContext *dc, TCGv src) |
| { |
| TCGv_ptr r_tickptr = tcg_temp_new_ptr(); |
| |
| tcg_gen_st_tl(src, tcg_env, env64_field_offsetof(hstick_cmpr)); |
| tcg_gen_ld_ptr(r_tickptr, tcg_env, env64_field_offsetof(hstick)); |
| translator_io_start(&dc->base); |
| gen_helper_tick_set_limit(r_tickptr, src); |
| /* End TB to handle timer interrupt */ |
| dc->base.is_jmp = DISAS_EXIT; |
| } |
| |
| TRANS(WRHPR_hstick_cmpr, HYPV, do_wr_special, a, hypervisor(dc), |
| do_wrhstick_cmpr) |
| |
| static bool do_saved_restored(DisasContext *dc, bool saved) |
| { |
| if (!supervisor(dc)) { |
| return raise_priv(dc); |
| } |
| if (saved) { |
| gen_helper_saved(tcg_env); |
| } else { |
| gen_helper_restored(tcg_env); |
| } |
| return advance_pc(dc); |
| } |
| |
| TRANS(SAVED, 64, do_saved_restored, true) |
| TRANS(RESTORED, 64, do_saved_restored, false) |
| |
| static bool trans_NOP(DisasContext *dc, arg_NOP *a) |
| { |
| return advance_pc(dc); |
| } |
| |
| /* |
| * TODO: Need a feature bit for sparcv8. |
| * In the meantime, treat all 32-bit cpus like sparcv7. |
| */ |
| TRANS(NOP_v7, 32, trans_NOP, a) |
| TRANS(NOP_v9, 64, trans_NOP, a) |
| |
| static bool do_arith_int(DisasContext *dc, arg_r_r_ri_cc *a, |
| void (*func)(TCGv, TCGv, TCGv), |
| void (*funci)(TCGv, TCGv, target_long), |
| bool logic_cc) |
| { |
| TCGv dst, src1; |
| |
| /* For simplicity, we under-decoded the rs2 form. */ |
| if (!a->imm && a->rs2_or_imm & ~0x1f) { |
| return false; |
| } |
| |
| if (logic_cc) { |
| dst = cpu_cc_N; |
| } else { |
| dst = gen_dest_gpr(dc, a->rd); |
| } |
| src1 = gen_load_gpr(dc, a->rs1); |
| |
| if (a->imm || a->rs2_or_imm == 0) { |
| if (funci) { |
| funci(dst, src1, a->rs2_or_imm); |
| } else { |
| func(dst, src1, tcg_constant_tl(a->rs2_or_imm)); |
| } |
| } else { |
| func(dst, src1, cpu_regs[a->rs2_or_imm]); |
| } |
| |
| if (logic_cc) { |
| if (TARGET_LONG_BITS == 64) { |
| tcg_gen_mov_tl(cpu_icc_Z, cpu_cc_N); |
| tcg_gen_movi_tl(cpu_icc_C, 0); |
| } |
| tcg_gen_mov_tl(cpu_cc_Z, cpu_cc_N); |
| tcg_gen_movi_tl(cpu_cc_C, 0); |
| tcg_gen_movi_tl(cpu_cc_V, 0); |
| } |
| |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool do_arith(DisasContext *dc, arg_r_r_ri_cc *a, |
| void (*func)(TCGv, TCGv, TCGv), |
| void (*funci)(TCGv, TCGv, target_long), |
| void (*func_cc)(TCGv, TCGv, TCGv)) |
| { |
| if (a->cc) { |
| return do_arith_int(dc, a, func_cc, NULL, false); |
| } |
| return do_arith_int(dc, a, func, funci, false); |
| } |
| |
| static bool do_logic(DisasContext *dc, arg_r_r_ri_cc *a, |
| void (*func)(TCGv, TCGv, TCGv), |
| void (*funci)(TCGv, TCGv, target_long)) |
| { |
| return do_arith_int(dc, a, func, funci, a->cc); |
| } |
| |
| TRANS(ADD, ALL, do_arith, a, tcg_gen_add_tl, tcg_gen_addi_tl, gen_op_addcc) |
| TRANS(SUB, ALL, do_arith, a, tcg_gen_sub_tl, tcg_gen_subi_tl, gen_op_subcc) |
| TRANS(ADDC, ALL, do_arith, a, gen_op_addc, NULL, gen_op_addccc) |
| TRANS(SUBC, ALL, do_arith, a, gen_op_subc, NULL, gen_op_subccc) |
| |
| TRANS(TADDcc, ALL, do_arith, a, NULL, NULL, gen_op_taddcc) |
| TRANS(TSUBcc, ALL, do_arith, a, NULL, NULL, gen_op_tsubcc) |
| TRANS(TADDccTV, ALL, do_arith, a, NULL, NULL, gen_op_taddcctv) |
| TRANS(TSUBccTV, ALL, do_arith, a, NULL, NULL, gen_op_tsubcctv) |
| |
| TRANS(AND, ALL, do_logic, a, tcg_gen_and_tl, tcg_gen_andi_tl) |
| TRANS(XOR, ALL, do_logic, a, tcg_gen_xor_tl, tcg_gen_xori_tl) |
| TRANS(ANDN, ALL, do_logic, a, tcg_gen_andc_tl, NULL) |
| TRANS(ORN, ALL, do_logic, a, tcg_gen_orc_tl, NULL) |
| TRANS(XORN, ALL, do_logic, a, tcg_gen_eqv_tl, NULL) |
| |
| TRANS(MULX, 64, do_arith, a, tcg_gen_mul_tl, tcg_gen_muli_tl, NULL) |
| TRANS(UMUL, MUL, do_logic, a, gen_op_umul, NULL) |
| TRANS(SMUL, MUL, do_logic, a, gen_op_smul, NULL) |
| TRANS(MULScc, ALL, do_arith, a, NULL, NULL, gen_op_mulscc) |
| |
| TRANS(UDIVcc, DIV, do_arith, a, NULL, NULL, gen_op_udivcc) |
| TRANS(SDIV, DIV, do_arith, a, gen_op_sdiv, NULL, gen_op_sdivcc) |
| |
| /* TODO: Should have feature bit -- comes in with UltraSparc T2. */ |
| TRANS(POPC, 64, do_arith, a, gen_op_popc, NULL, NULL) |
| |
| static bool trans_OR(DisasContext *dc, arg_r_r_ri_cc *a) |
| { |
| /* OR with %g0 is the canonical alias for MOV. */ |
| if (!a->cc && a->rs1 == 0) { |
| if (a->imm || a->rs2_or_imm == 0) { |
| gen_store_gpr(dc, a->rd, tcg_constant_tl(a->rs2_or_imm)); |
| } else if (a->rs2_or_imm & ~0x1f) { |
| /* For simplicity, we under-decoded the rs2 form. */ |
| return false; |
| } else { |
| gen_store_gpr(dc, a->rd, cpu_regs[a->rs2_or_imm]); |
| } |
| return advance_pc(dc); |
| } |
| return do_logic(dc, a, tcg_gen_or_tl, tcg_gen_ori_tl); |
| } |
| |
| static bool trans_UDIV(DisasContext *dc, arg_r_r_ri *a) |
| { |
| TCGv_i64 t1, t2; |
| TCGv dst; |
| |
| if (!avail_DIV(dc)) { |
| return false; |
| } |
| /* For simplicity, we under-decoded the rs2 form. */ |
| if (!a->imm && a->rs2_or_imm & ~0x1f) { |
| return false; |
| } |
| |
| if (unlikely(a->rs2_or_imm == 0)) { |
| gen_exception(dc, TT_DIV_ZERO); |
| return true; |
| } |
| |
| if (a->imm) { |
| t2 = tcg_constant_i64((uint32_t)a->rs2_or_imm); |
| } else { |
| TCGLabel *lab; |
| TCGv_i32 n2; |
| |
| finishing_insn(dc); |
| flush_cond(dc); |
| |
| n2 = tcg_temp_new_i32(); |
| tcg_gen_trunc_tl_i32(n2, cpu_regs[a->rs2_or_imm]); |
| |
| lab = delay_exception(dc, TT_DIV_ZERO); |
| tcg_gen_brcondi_i32(TCG_COND_EQ, n2, 0, lab); |
| |
| t2 = tcg_temp_new_i64(); |
| #ifdef TARGET_SPARC64 |
| tcg_gen_ext32u_i64(t2, cpu_regs[a->rs2_or_imm]); |
| #else |
| tcg_gen_extu_i32_i64(t2, cpu_regs[a->rs2_or_imm]); |
| #endif |
| } |
| |
| t1 = tcg_temp_new_i64(); |
| tcg_gen_concat_tl_i64(t1, gen_load_gpr(dc, a->rs1), cpu_y); |
| |
| tcg_gen_divu_i64(t1, t1, t2); |
| tcg_gen_umin_i64(t1, t1, tcg_constant_i64(UINT32_MAX)); |
| |
| dst = gen_dest_gpr(dc, a->rd); |
| tcg_gen_trunc_i64_tl(dst, t1); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_UDIVX(DisasContext *dc, arg_r_r_ri *a) |
| { |
| TCGv dst, src1, src2; |
| |
| if (!avail_64(dc)) { |
| return false; |
| } |
| /* For simplicity, we under-decoded the rs2 form. */ |
| if (!a->imm && a->rs2_or_imm & ~0x1f) { |
| return false; |
| } |
| |
| if (unlikely(a->rs2_or_imm == 0)) { |
| gen_exception(dc, TT_DIV_ZERO); |
| return true; |
| } |
| |
| if (a->imm) { |
| src2 = tcg_constant_tl(a->rs2_or_imm); |
| } else { |
| TCGLabel *lab; |
| |
| finishing_insn(dc); |
| flush_cond(dc); |
| |
| lab = delay_exception(dc, TT_DIV_ZERO); |
| src2 = cpu_regs[a->rs2_or_imm]; |
| tcg_gen_brcondi_tl(TCG_COND_EQ, src2, 0, lab); |
| } |
| |
| dst = gen_dest_gpr(dc, a->rd); |
| src1 = gen_load_gpr(dc, a->rs1); |
| |
| tcg_gen_divu_tl(dst, src1, src2); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_SDIVX(DisasContext *dc, arg_r_r_ri *a) |
| { |
| TCGv dst, src1, src2; |
| |
| if (!avail_64(dc)) { |
| return false; |
| } |
| /* For simplicity, we under-decoded the rs2 form. */ |
| if (!a->imm && a->rs2_or_imm & ~0x1f) { |
| return false; |
| } |
| |
| if (unlikely(a->rs2_or_imm == 0)) { |
| gen_exception(dc, TT_DIV_ZERO); |
| return true; |
| } |
| |
| dst = gen_dest_gpr(dc, a->rd); |
| src1 = gen_load_gpr(dc, a->rs1); |
| |
| if (a->imm) { |
| if (unlikely(a->rs2_or_imm == -1)) { |
| tcg_gen_neg_tl(dst, src1); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| src2 = tcg_constant_tl(a->rs2_or_imm); |
| } else { |
| TCGLabel *lab; |
| TCGv t1, t2; |
| |
| finishing_insn(dc); |
| flush_cond(dc); |
| |
| lab = delay_exception(dc, TT_DIV_ZERO); |
| src2 = cpu_regs[a->rs2_or_imm]; |
| tcg_gen_brcondi_tl(TCG_COND_EQ, src2, 0, lab); |
| |
| /* |
| * Need to avoid INT64_MIN / -1, which will trap on x86 host. |
| * Set SRC2 to 1 as a new divisor, to produce the correct result. |
| */ |
| t1 = tcg_temp_new(); |
| t2 = tcg_temp_new(); |
| tcg_gen_setcondi_tl(TCG_COND_EQ, t1, src1, (target_long)INT64_MIN); |
| tcg_gen_setcondi_tl(TCG_COND_EQ, t2, src2, -1); |
| tcg_gen_and_tl(t1, t1, t2); |
| tcg_gen_movcond_tl(TCG_COND_NE, t1, t1, tcg_constant_tl(0), |
| tcg_constant_tl(1), src2); |
| src2 = t1; |
| } |
| |
| tcg_gen_div_tl(dst, src1, src2); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool gen_edge(DisasContext *dc, arg_r_r_r *a, |
| int width, bool cc, bool little_endian) |
| { |
| TCGv dst, s1, s2, l, r, t, m; |
| uint64_t amask = address_mask_i(dc, -8); |
| |
| dst = gen_dest_gpr(dc, a->rd); |
| s1 = gen_load_gpr(dc, a->rs1); |
| s2 = gen_load_gpr(dc, a->rs2); |
| |
| if (cc) { |
| gen_op_subcc(cpu_cc_N, s1, s2); |
| } |
| |
| l = tcg_temp_new(); |
| r = tcg_temp_new(); |
| t = tcg_temp_new(); |
| |
| switch (width) { |
| case 8: |
| tcg_gen_andi_tl(l, s1, 7); |
| tcg_gen_andi_tl(r, s2, 7); |
| tcg_gen_xori_tl(r, r, 7); |
| m = tcg_constant_tl(0xff); |
| break; |
| case 16: |
| tcg_gen_extract_tl(l, s1, 1, 2); |
| tcg_gen_extract_tl(r, s2, 1, 2); |
| tcg_gen_xori_tl(r, r, 3); |
| m = tcg_constant_tl(0xf); |
| break; |
| case 32: |
| tcg_gen_extract_tl(l, s1, 2, 1); |
| tcg_gen_extract_tl(r, s2, 2, 1); |
| tcg_gen_xori_tl(r, r, 1); |
| m = tcg_constant_tl(0x3); |
| break; |
| default: |
| abort(); |
| } |
| |
| /* Compute Left Edge */ |
| if (little_endian) { |
| tcg_gen_shl_tl(l, m, l); |
| tcg_gen_and_tl(l, l, m); |
| } else { |
| tcg_gen_shr_tl(l, m, l); |
| } |
| /* Compute Right Edge */ |
| if (little_endian) { |
| tcg_gen_shr_tl(r, m, r); |
| } else { |
| tcg_gen_shl_tl(r, m, r); |
| tcg_gen_and_tl(r, r, m); |
| } |
| |
| /* Compute dst = (s1 == s2 under amask ? l : l & r) */ |
| tcg_gen_xor_tl(t, s1, s2); |
| tcg_gen_and_tl(r, r, l); |
| tcg_gen_movcond_tl(TCG_COND_TSTEQ, dst, t, tcg_constant_tl(amask), r, l); |
| |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(EDGE8cc, VIS1, gen_edge, a, 8, 1, 0) |
| TRANS(EDGE8Lcc, VIS1, gen_edge, a, 8, 1, 1) |
| TRANS(EDGE16cc, VIS1, gen_edge, a, 16, 1, 0) |
| TRANS(EDGE16Lcc, VIS1, gen_edge, a, 16, 1, 1) |
| TRANS(EDGE32cc, VIS1, gen_edge, a, 32, 1, 0) |
| TRANS(EDGE32Lcc, VIS1, gen_edge, a, 32, 1, 1) |
| |
| TRANS(EDGE8N, VIS2, gen_edge, a, 8, 0, 0) |
| TRANS(EDGE8LN, VIS2, gen_edge, a, 8, 0, 1) |
| TRANS(EDGE16N, VIS2, gen_edge, a, 16, 0, 0) |
| TRANS(EDGE16LN, VIS2, gen_edge, a, 16, 0, 1) |
| TRANS(EDGE32N, VIS2, gen_edge, a, 32, 0, 0) |
| TRANS(EDGE32LN, VIS2, gen_edge, a, 32, 0, 1) |
| |
| static bool do_rr(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv, TCGv)) |
| { |
| TCGv dst = gen_dest_gpr(dc, a->rd); |
| TCGv src = gen_load_gpr(dc, a->rs); |
| |
| func(dst, src); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(LZCNT, VIS3, do_rr, a, gen_op_lzcnt) |
| |
| static bool do_rrr(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv, TCGv, TCGv)) |
| { |
| TCGv dst = gen_dest_gpr(dc, a->rd); |
| TCGv src1 = gen_load_gpr(dc, a->rs1); |
| TCGv src2 = gen_load_gpr(dc, a->rs2); |
| |
| func(dst, src1, src2); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(ARRAY8, VIS1, do_rrr, a, gen_helper_array8) |
| TRANS(ARRAY16, VIS1, do_rrr, a, gen_op_array16) |
| TRANS(ARRAY32, VIS1, do_rrr, a, gen_op_array32) |
| |
| TRANS(ADDXC, VIS3, do_rrr, a, gen_op_addxc) |
| TRANS(ADDXCcc, VIS3, do_rrr, a, gen_op_addxccc) |
| |
| TRANS(SUBXC, VIS4, do_rrr, a, gen_op_subxc) |
| TRANS(SUBXCcc, VIS4, do_rrr, a, gen_op_subxccc) |
| |
| TRANS(UMULXHI, VIS3, do_rrr, a, gen_op_umulxhi) |
| |
| static void gen_op_alignaddr(TCGv dst, TCGv s1, TCGv s2) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv tmp = tcg_temp_new(); |
| |
| tcg_gen_add_tl(tmp, s1, s2); |
| tcg_gen_andi_tl(dst, tmp, -8); |
| tcg_gen_deposit_tl(cpu_gsr, cpu_gsr, tmp, 0, 3); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| static void gen_op_alignaddrl(TCGv dst, TCGv s1, TCGv s2) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv tmp = tcg_temp_new(); |
| |
| tcg_gen_add_tl(tmp, s1, s2); |
| tcg_gen_andi_tl(dst, tmp, -8); |
| tcg_gen_neg_tl(tmp, tmp); |
| tcg_gen_deposit_tl(cpu_gsr, cpu_gsr, tmp, 0, 3); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| TRANS(ALIGNADDR, VIS1, do_rrr, a, gen_op_alignaddr) |
| TRANS(ALIGNADDRL, VIS1, do_rrr, a, gen_op_alignaddrl) |
| |
| static void gen_op_bmask(TCGv dst, TCGv s1, TCGv s2) |
| { |
| #ifdef TARGET_SPARC64 |
| tcg_gen_add_tl(dst, s1, s2); |
| tcg_gen_deposit_tl(cpu_gsr, cpu_gsr, dst, 32, 32); |
| #else |
| g_assert_not_reached(); |
| #endif |
| } |
| |
| TRANS(BMASK, VIS2, do_rrr, a, gen_op_bmask) |
| |
| static bool do_cmask(DisasContext *dc, int rs2, void (*func)(TCGv, TCGv, TCGv)) |
| { |
| func(cpu_gsr, cpu_gsr, gen_load_gpr(dc, rs2)); |
| return true; |
| } |
| |
| TRANS(CMASK8, VIS3, do_cmask, a->rs2, gen_helper_cmask8) |
| TRANS(CMASK16, VIS3, do_cmask, a->rs2, gen_helper_cmask16) |
| TRANS(CMASK32, VIS3, do_cmask, a->rs2, gen_helper_cmask32) |
| |
| static bool do_shift_r(DisasContext *dc, arg_shiftr *a, bool l, bool u) |
| { |
| TCGv dst, src1, src2; |
| |
| /* Reject 64-bit shifts for sparc32. */ |
| if (avail_32(dc) && a->x) { |
| return false; |
| } |
| |
| src2 = tcg_temp_new(); |
| tcg_gen_andi_tl(src2, gen_load_gpr(dc, a->rs2), a->x ? 63 : 31); |
| src1 = gen_load_gpr(dc, a->rs1); |
| dst = gen_dest_gpr(dc, a->rd); |
| |
| if (l) { |
| tcg_gen_shl_tl(dst, src1, src2); |
| if (!a->x) { |
| tcg_gen_ext32u_tl(dst, dst); |
| } |
| } else if (u) { |
| if (!a->x) { |
| tcg_gen_ext32u_tl(dst, src1); |
| src1 = dst; |
| } |
| tcg_gen_shr_tl(dst, src1, src2); |
| } else { |
| if (!a->x) { |
| tcg_gen_ext32s_tl(dst, src1); |
| src1 = dst; |
| } |
| tcg_gen_sar_tl(dst, src1, src2); |
| } |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(SLL_r, ALL, do_shift_r, a, true, true) |
| TRANS(SRL_r, ALL, do_shift_r, a, false, true) |
| TRANS(SRA_r, ALL, do_shift_r, a, false, false) |
| |
| static bool do_shift_i(DisasContext *dc, arg_shifti *a, bool l, bool u) |
| { |
| TCGv dst, src1; |
| |
| /* Reject 64-bit shifts for sparc32. */ |
| if (avail_32(dc) && (a->x || a->i >= 32)) { |
| return false; |
| } |
| |
| src1 = gen_load_gpr(dc, a->rs1); |
| dst = gen_dest_gpr(dc, a->rd); |
| |
| if (avail_32(dc) || a->x) { |
| if (l) { |
| tcg_gen_shli_tl(dst, src1, a->i); |
| } else if (u) { |
| tcg_gen_shri_tl(dst, src1, a->i); |
| } else { |
| tcg_gen_sari_tl(dst, src1, a->i); |
| } |
| } else { |
| if (l) { |
| tcg_gen_deposit_z_tl(dst, src1, a->i, 32 - a->i); |
| } else if (u) { |
| tcg_gen_extract_tl(dst, src1, a->i, 32 - a->i); |
| } else { |
| tcg_gen_sextract_tl(dst, src1, a->i, 32 - a->i); |
| } |
| } |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(SLL_i, ALL, do_shift_i, a, true, true) |
| TRANS(SRL_i, ALL, do_shift_i, a, false, true) |
| TRANS(SRA_i, ALL, do_shift_i, a, false, false) |
| |
| static TCGv gen_rs2_or_imm(DisasContext *dc, bool imm, int rs2_or_imm) |
| { |
| /* For simplicity, we under-decoded the rs2 form. */ |
| if (!imm && rs2_or_imm & ~0x1f) { |
| return NULL; |
| } |
| if (imm || rs2_or_imm == 0) { |
| return tcg_constant_tl(rs2_or_imm); |
| } else { |
| return cpu_regs[rs2_or_imm]; |
| } |
| } |
| |
| static bool do_mov_cond(DisasContext *dc, DisasCompare *cmp, int rd, TCGv src2) |
| { |
| TCGv dst = gen_load_gpr(dc, rd); |
| TCGv c2 = tcg_constant_tl(cmp->c2); |
| |
| tcg_gen_movcond_tl(cmp->cond, dst, cmp->c1, c2, src2, dst); |
| gen_store_gpr(dc, rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_MOVcc(DisasContext *dc, arg_MOVcc *a) |
| { |
| TCGv src2 = gen_rs2_or_imm(dc, a->imm, a->rs2_or_imm); |
| DisasCompare cmp; |
| |
| if (src2 == NULL) { |
| return false; |
| } |
| gen_compare(&cmp, a->cc, a->cond, dc); |
| return do_mov_cond(dc, &cmp, a->rd, src2); |
| } |
| |
| static bool trans_MOVfcc(DisasContext *dc, arg_MOVfcc *a) |
| { |
| TCGv src2 = gen_rs2_or_imm(dc, a->imm, a->rs2_or_imm); |
| DisasCompare cmp; |
| |
| if (src2 == NULL) { |
| return false; |
| } |
| gen_fcompare(&cmp, a->cc, a->cond); |
| return do_mov_cond(dc, &cmp, a->rd, src2); |
| } |
| |
| static bool trans_MOVR(DisasContext *dc, arg_MOVR *a) |
| { |
| TCGv src2 = gen_rs2_or_imm(dc, a->imm, a->rs2_or_imm); |
| DisasCompare cmp; |
| |
| if (src2 == NULL) { |
| return false; |
| } |
| if (!gen_compare_reg(&cmp, a->cond, gen_load_gpr(dc, a->rs1))) { |
| return false; |
| } |
| return do_mov_cond(dc, &cmp, a->rd, src2); |
| } |
| |
| static bool do_add_special(DisasContext *dc, arg_r_r_ri *a, |
| bool (*func)(DisasContext *dc, int rd, TCGv src)) |
| { |
| TCGv src1, sum; |
| |
| /* For simplicity, we under-decoded the rs2 form. */ |
| if (!a->imm && a->rs2_or_imm & ~0x1f) { |
| return false; |
| } |
| |
| /* |
| * Always load the sum into a new temporary. |
| * This is required to capture the value across a window change, |
| * e.g. SAVE and RESTORE, and may be optimized away otherwise. |
| */ |
| sum = tcg_temp_new(); |
| src1 = gen_load_gpr(dc, a->rs1); |
| if (a->imm || a->rs2_or_imm == 0) { |
| tcg_gen_addi_tl(sum, src1, a->rs2_or_imm); |
| } else { |
| tcg_gen_add_tl(sum, src1, cpu_regs[a->rs2_or_imm]); |
| } |
| return func(dc, a->rd, sum); |
| } |
| |
| static bool do_jmpl(DisasContext *dc, int rd, TCGv src) |
| { |
| /* |
| * Preserve pc across advance, so that we can delay |
| * the writeback to rd until after src is consumed. |
| */ |
| target_ulong cur_pc = dc->pc; |
| |
| gen_check_align(dc, src, 3); |
| |
| gen_mov_pc_npc(dc); |
| tcg_gen_mov_tl(cpu_npc, src); |
| gen_address_mask(dc, cpu_npc); |
| gen_store_gpr(dc, rd, tcg_constant_tl(cur_pc)); |
| |
| dc->npc = DYNAMIC_PC_LOOKUP; |
| return true; |
| } |
| |
| TRANS(JMPL, ALL, do_add_special, a, do_jmpl) |
| |
| static bool do_rett(DisasContext *dc, int rd, TCGv src) |
| { |
| if (!supervisor(dc)) { |
| return raise_priv(dc); |
| } |
| |
| gen_check_align(dc, src, 3); |
| |
| gen_mov_pc_npc(dc); |
| tcg_gen_mov_tl(cpu_npc, src); |
| gen_helper_rett(tcg_env); |
| |
| dc->npc = DYNAMIC_PC; |
| return true; |
| } |
| |
| TRANS(RETT, 32, do_add_special, a, do_rett) |
| |
| static bool do_return(DisasContext *dc, int rd, TCGv src) |
| { |
| gen_check_align(dc, src, 3); |
| gen_helper_restore(tcg_env); |
| |
| gen_mov_pc_npc(dc); |
| tcg_gen_mov_tl(cpu_npc, src); |
| gen_address_mask(dc, cpu_npc); |
| |
| dc->npc = DYNAMIC_PC_LOOKUP; |
| return true; |
| } |
| |
| TRANS(RETURN, 64, do_add_special, a, do_return) |
| |
| static bool do_save(DisasContext *dc, int rd, TCGv src) |
| { |
| gen_helper_save(tcg_env); |
| gen_store_gpr(dc, rd, src); |
| return advance_pc(dc); |
| } |
| |
| TRANS(SAVE, ALL, do_add_special, a, do_save) |
| |
| static bool do_restore(DisasContext *dc, int rd, TCGv src) |
| { |
| gen_helper_restore(tcg_env); |
| gen_store_gpr(dc, rd, src); |
| return advance_pc(dc); |
| } |
| |
| TRANS(RESTORE, ALL, do_add_special, a, do_restore) |
| |
| static bool do_done_retry(DisasContext *dc, bool done) |
| { |
| if (!supervisor(dc)) { |
| return raise_priv(dc); |
| } |
| dc->npc = DYNAMIC_PC; |
| dc->pc = DYNAMIC_PC; |
| translator_io_start(&dc->base); |
| if (done) { |
| gen_helper_done(tcg_env); |
| } else { |
| gen_helper_retry(tcg_env); |
| } |
| return true; |
| } |
| |
| TRANS(DONE, 64, do_done_retry, true) |
| TRANS(RETRY, 64, do_done_retry, false) |
| |
| /* |
| * Major opcode 11 -- load and store instructions |
| */ |
| |
| static TCGv gen_ldst_addr(DisasContext *dc, int rs1, bool imm, int rs2_or_imm) |
| { |
| TCGv addr, tmp = NULL; |
| |
| /* For simplicity, we under-decoded the rs2 form. */ |
| if (!imm && rs2_or_imm & ~0x1f) { |
| return NULL; |
| } |
| |
| addr = gen_load_gpr(dc, rs1); |
| if (rs2_or_imm) { |
| tmp = tcg_temp_new(); |
| if (imm) { |
| tcg_gen_addi_tl(tmp, addr, rs2_or_imm); |
| } else { |
| tcg_gen_add_tl(tmp, addr, cpu_regs[rs2_or_imm]); |
| } |
| addr = tmp; |
| } |
| if (AM_CHECK(dc)) { |
| if (!tmp) { |
| tmp = tcg_temp_new(); |
| } |
| tcg_gen_ext32u_tl(tmp, addr); |
| addr = tmp; |
| } |
| return addr; |
| } |
| |
| static bool do_ld_gpr(DisasContext *dc, arg_r_r_ri_asi *a, MemOp mop) |
| { |
| TCGv reg, addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| DisasASI da; |
| |
| if (addr == NULL) { |
| return false; |
| } |
| da = resolve_asi(dc, a->asi, mop); |
| |
| reg = gen_dest_gpr(dc, a->rd); |
| gen_ld_asi(dc, &da, reg, addr); |
| gen_store_gpr(dc, a->rd, reg); |
| return advance_pc(dc); |
| } |
| |
| TRANS(LDUW, ALL, do_ld_gpr, a, MO_TEUL) |
| TRANS(LDUB, ALL, do_ld_gpr, a, MO_UB) |
| TRANS(LDUH, ALL, do_ld_gpr, a, MO_TEUW) |
| TRANS(LDSB, ALL, do_ld_gpr, a, MO_SB) |
| TRANS(LDSH, ALL, do_ld_gpr, a, MO_TESW) |
| TRANS(LDSW, 64, do_ld_gpr, a, MO_TESL) |
| TRANS(LDX, 64, do_ld_gpr, a, MO_TEUQ) |
| |
| static bool do_st_gpr(DisasContext *dc, arg_r_r_ri_asi *a, MemOp mop) |
| { |
| TCGv reg, addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| DisasASI da; |
| |
| if (addr == NULL) { |
| return false; |
| } |
| da = resolve_asi(dc, a->asi, mop); |
| |
| reg = gen_load_gpr(dc, a->rd); |
| gen_st_asi(dc, &da, reg, addr); |
| return advance_pc(dc); |
| } |
| |
| TRANS(STW, ALL, do_st_gpr, a, MO_TEUL) |
| TRANS(STB, ALL, do_st_gpr, a, MO_UB) |
| TRANS(STH, ALL, do_st_gpr, a, MO_TEUW) |
| TRANS(STX, 64, do_st_gpr, a, MO_TEUQ) |
| |
| static bool trans_LDD(DisasContext *dc, arg_r_r_ri_asi *a) |
| { |
| TCGv addr; |
| DisasASI da; |
| |
| if (a->rd & 1) { |
| return false; |
| } |
| addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| if (addr == NULL) { |
| return false; |
| } |
| da = resolve_asi(dc, a->asi, MO_TEUQ); |
| gen_ldda_asi(dc, &da, addr, a->rd); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_STD(DisasContext *dc, arg_r_r_ri_asi *a) |
| { |
| TCGv addr; |
| DisasASI da; |
| |
| if (a->rd & 1) { |
| return false; |
| } |
| addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| if (addr == NULL) { |
| return false; |
| } |
| da = resolve_asi(dc, a->asi, MO_TEUQ); |
| gen_stda_asi(dc, &da, addr, a->rd); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_LDSTUB(DisasContext *dc, arg_r_r_ri_asi *a) |
| { |
| TCGv addr, reg; |
| DisasASI da; |
| |
| addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| if (addr == NULL) { |
| return false; |
| } |
| da = resolve_asi(dc, a->asi, MO_UB); |
| |
| reg = gen_dest_gpr(dc, a->rd); |
| gen_ldstub_asi(dc, &da, reg, addr); |
| gen_store_gpr(dc, a->rd, reg); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_SWAP(DisasContext *dc, arg_r_r_ri_asi *a) |
| { |
| TCGv addr, dst, src; |
| DisasASI da; |
| |
| addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| if (addr == NULL) { |
| return false; |
| } |
| da = resolve_asi(dc, a->asi, MO_TEUL); |
| |
| dst = gen_dest_gpr(dc, a->rd); |
| src = gen_load_gpr(dc, a->rd); |
| gen_swap_asi(dc, &da, dst, src, addr); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool do_casa(DisasContext *dc, arg_r_r_ri_asi *a, MemOp mop) |
| { |
| TCGv addr, o, n, c; |
| DisasASI da; |
| |
| addr = gen_ldst_addr(dc, a->rs1, true, 0); |
| if (addr == NULL) { |
| return false; |
| } |
| da = resolve_asi(dc, a->asi, mop); |
| |
| o = gen_dest_gpr(dc, a->rd); |
| n = gen_load_gpr(dc, a->rd); |
| c = gen_load_gpr(dc, a->rs2_or_imm); |
| gen_cas_asi(dc, &da, o, n, c, addr); |
| gen_store_gpr(dc, a->rd, o); |
| return advance_pc(dc); |
| } |
| |
| TRANS(CASA, CASA, do_casa, a, MO_TEUL) |
| TRANS(CASXA, 64, do_casa, a, MO_TEUQ) |
| |
| static bool do_ld_fpr(DisasContext *dc, arg_r_r_ri_asi *a, MemOp sz) |
| { |
| TCGv addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| DisasASI da; |
| |
| if (addr == NULL) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (sz == MO_128 && gen_trap_float128(dc)) { |
| return true; |
| } |
| da = resolve_asi(dc, a->asi, MO_TE | sz); |
| gen_ldf_asi(dc, &da, sz, addr, a->rd); |
| gen_update_fprs_dirty(dc, a->rd); |
| return advance_pc(dc); |
| } |
| |
| TRANS(LDF, ALL, do_ld_fpr, a, MO_32) |
| TRANS(LDDF, ALL, do_ld_fpr, a, MO_64) |
| TRANS(LDQF, ALL, do_ld_fpr, a, MO_128) |
| |
| TRANS(LDFA, 64, do_ld_fpr, a, MO_32) |
| TRANS(LDDFA, 64, do_ld_fpr, a, MO_64) |
| TRANS(LDQFA, 64, do_ld_fpr, a, MO_128) |
| |
| static bool do_st_fpr(DisasContext *dc, arg_r_r_ri_asi *a, MemOp sz) |
| { |
| TCGv addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| DisasASI da; |
| |
| if (addr == NULL) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (sz == MO_128 && gen_trap_float128(dc)) { |
| return true; |
| } |
| da = resolve_asi(dc, a->asi, MO_TE | sz); |
| gen_stf_asi(dc, &da, sz, addr, a->rd); |
| return advance_pc(dc); |
| } |
| |
| TRANS(STF, ALL, do_st_fpr, a, MO_32) |
| TRANS(STDF, ALL, do_st_fpr, a, MO_64) |
| TRANS(STQF, 64, do_st_fpr, a, MO_128) |
| |
| TRANS(STFA, 64, do_st_fpr, a, MO_32) |
| TRANS(STDFA, 64, do_st_fpr, a, MO_64) |
| TRANS(STQFA, 64, do_st_fpr, a, MO_128) |
| |
| static bool trans_STDFQ(DisasContext *dc, arg_STDFQ *a) |
| { |
| if (!avail_32(dc)) { |
| return false; |
| } |
| if (!supervisor(dc)) { |
| return raise_priv(dc); |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| gen_op_fpexception_im(dc, FSR_FTT_SEQ_ERROR); |
| return true; |
| } |
| |
| static bool trans_LDFSR(DisasContext *dc, arg_r_r_ri *a) |
| { |
| TCGv addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| TCGv_i32 tmp; |
| |
| if (addr == NULL) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| tmp = tcg_temp_new_i32(); |
| tcg_gen_qemu_ld_i32(tmp, addr, dc->mem_idx, MO_TEUL | MO_ALIGN); |
| |
| tcg_gen_extract_i32(cpu_fcc[0], tmp, FSR_FCC0_SHIFT, 2); |
| /* LDFSR does not change FCC[1-3]. */ |
| |
| gen_helper_set_fsr_nofcc_noftt(tcg_env, tmp); |
| return advance_pc(dc); |
| } |
| |
| static bool do_ldxfsr(DisasContext *dc, arg_r_r_ri *a, bool entire) |
| { |
| #ifdef TARGET_SPARC64 |
| TCGv addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| TCGv_i64 t64; |
| TCGv_i32 lo, hi; |
| |
| if (addr == NULL) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| t64 = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(t64, addr, dc->mem_idx, MO_TEUQ | MO_ALIGN); |
| |
| lo = tcg_temp_new_i32(); |
| hi = cpu_fcc[3]; |
| tcg_gen_extr_i64_i32(lo, hi, t64); |
| tcg_gen_extract_i32(cpu_fcc[0], lo, FSR_FCC0_SHIFT, 2); |
| tcg_gen_extract_i32(cpu_fcc[1], hi, FSR_FCC1_SHIFT - 32, 2); |
| tcg_gen_extract_i32(cpu_fcc[2], hi, FSR_FCC2_SHIFT - 32, 2); |
| tcg_gen_extract_i32(cpu_fcc[3], hi, FSR_FCC3_SHIFT - 32, 2); |
| |
| if (entire) { |
| gen_helper_set_fsr_nofcc(tcg_env, lo); |
| } else { |
| gen_helper_set_fsr_nofcc_noftt(tcg_env, lo); |
| } |
| return advance_pc(dc); |
| #else |
| return false; |
| #endif |
| } |
| |
| TRANS(LDXFSR, 64, do_ldxfsr, a, false) |
| TRANS(LDXEFSR, VIS3B, do_ldxfsr, a, true) |
| |
| static bool do_stfsr(DisasContext *dc, arg_r_r_ri *a, MemOp mop) |
| { |
| TCGv addr = gen_ldst_addr(dc, a->rs1, a->imm, a->rs2_or_imm); |
| TCGv fsr; |
| |
| if (addr == NULL) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| fsr = tcg_temp_new(); |
| gen_helper_get_fsr(fsr, tcg_env); |
| tcg_gen_qemu_st_tl(fsr, addr, dc->mem_idx, mop | MO_ALIGN); |
| return advance_pc(dc); |
| } |
| |
| TRANS(STFSR, ALL, do_stfsr, a, MO_TEUL) |
| TRANS(STXFSR, 64, do_stfsr, a, MO_TEUQ) |
| |
| static bool do_fc(DisasContext *dc, int rd, int32_t c) |
| { |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| gen_store_fpr_F(dc, rd, tcg_constant_i32(c)); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FZEROs, VIS1, do_fc, a->rd, 0) |
| TRANS(FONEs, VIS1, do_fc, a->rd, -1) |
| |
| static bool do_dc(DisasContext *dc, int rd, int64_t c) |
| { |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| gen_store_fpr_D(dc, rd, tcg_constant_i64(c)); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FZEROd, VIS1, do_dc, a->rd, 0) |
| TRANS(FONEd, VIS1, do_dc, a->rd, -1) |
| |
| static bool do_ff(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i32, TCGv_i32)) |
| { |
| TCGv_i32 tmp; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| tmp = gen_load_fpr_F(dc, a->rs); |
| func(tmp, tmp); |
| gen_store_fpr_F(dc, a->rd, tmp); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMOVs, ALL, do_ff, a, gen_op_fmovs) |
| TRANS(FNEGs, ALL, do_ff, a, gen_op_fnegs) |
| TRANS(FABSs, ALL, do_ff, a, gen_op_fabss) |
| TRANS(FSRCs, VIS1, do_ff, a, tcg_gen_mov_i32) |
| TRANS(FNOTs, VIS1, do_ff, a, tcg_gen_not_i32) |
| |
| static bool do_fd(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i32, TCGv_i64)) |
| { |
| TCGv_i32 dst; |
| TCGv_i64 src; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i32(); |
| src = gen_load_fpr_D(dc, a->rs); |
| func(dst, src); |
| gen_store_fpr_F(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FPACK16, VIS1, do_fd, a, gen_op_fpack16) |
| TRANS(FPACKFIX, VIS1, do_fd, a, gen_op_fpackfix) |
| |
| static bool do_env_ff(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i32, TCGv_env, TCGv_i32)) |
| { |
| TCGv_i32 tmp; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| tmp = gen_load_fpr_F(dc, a->rs); |
| func(tmp, tcg_env, tmp); |
| gen_store_fpr_F(dc, a->rd, tmp); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FSQRTs, ALL, do_env_ff, a, gen_helper_fsqrts) |
| TRANS(FiTOs, ALL, do_env_ff, a, gen_helper_fitos) |
| TRANS(FsTOi, ALL, do_env_ff, a, gen_helper_fstoi) |
| |
| static bool do_env_fd(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i32, TCGv_env, TCGv_i64)) |
| { |
| TCGv_i32 dst; |
| TCGv_i64 src; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i32(); |
| src = gen_load_fpr_D(dc, a->rs); |
| func(dst, tcg_env, src); |
| gen_store_fpr_F(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FdTOs, ALL, do_env_fd, a, gen_helper_fdtos) |
| TRANS(FdTOi, ALL, do_env_fd, a, gen_helper_fdtoi) |
| TRANS(FxTOs, 64, do_env_fd, a, gen_helper_fxtos) |
| |
| static bool do_dd(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i64, TCGv_i64)) |
| { |
| TCGv_i64 dst, src; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src = gen_load_fpr_D(dc, a->rs); |
| func(dst, src); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMOVd, 64, do_dd, a, gen_op_fmovd) |
| TRANS(FNEGd, 64, do_dd, a, gen_op_fnegd) |
| TRANS(FABSd, 64, do_dd, a, gen_op_fabsd) |
| TRANS(FSRCd, VIS1, do_dd, a, tcg_gen_mov_i64) |
| TRANS(FNOTd, VIS1, do_dd, a, tcg_gen_not_i64) |
| |
| static bool do_env_dd(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i64, TCGv_env, TCGv_i64)) |
| { |
| TCGv_i64 dst, src; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src = gen_load_fpr_D(dc, a->rs); |
| func(dst, tcg_env, src); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FSQRTd, ALL, do_env_dd, a, gen_helper_fsqrtd) |
| TRANS(FxTOd, 64, do_env_dd, a, gen_helper_fxtod) |
| TRANS(FdTOx, 64, do_env_dd, a, gen_helper_fdtox) |
| |
| static bool do_df(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i64, TCGv_i32)) |
| { |
| TCGv_i64 dst; |
| TCGv_i32 src; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src = gen_load_fpr_F(dc, a->rs); |
| func(dst, src); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FEXPAND, VIS1, do_df, a, gen_helper_fexpand) |
| |
| static bool do_env_df(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i64, TCGv_env, TCGv_i32)) |
| { |
| TCGv_i64 dst; |
| TCGv_i32 src; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src = gen_load_fpr_F(dc, a->rs); |
| func(dst, tcg_env, src); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FiTOd, ALL, do_env_df, a, gen_helper_fitod) |
| TRANS(FsTOd, ALL, do_env_df, a, gen_helper_fstod) |
| TRANS(FsTOx, 64, do_env_df, a, gen_helper_fstox) |
| |
| static bool do_qq(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i128, TCGv_i128)) |
| { |
| TCGv_i128 t; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| gen_op_clear_ieee_excp_and_FTT(); |
| t = gen_load_fpr_Q(dc, a->rs); |
| func(t, t); |
| gen_store_fpr_Q(dc, a->rd, t); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMOVq, 64, do_qq, a, tcg_gen_mov_i128) |
| TRANS(FNEGq, 64, do_qq, a, gen_op_fnegq) |
| TRANS(FABSq, 64, do_qq, a, gen_op_fabsq) |
| |
| static bool do_env_qq(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i128, TCGv_env, TCGv_i128)) |
| { |
| TCGv_i128 t; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| t = gen_load_fpr_Q(dc, a->rs); |
| func(t, tcg_env, t); |
| gen_store_fpr_Q(dc, a->rd, t); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FSQRTq, ALL, do_env_qq, a, gen_helper_fsqrtq) |
| |
| static bool do_env_fq(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i32, TCGv_env, TCGv_i128)) |
| { |
| TCGv_i128 src; |
| TCGv_i32 dst; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| src = gen_load_fpr_Q(dc, a->rs); |
| dst = tcg_temp_new_i32(); |
| func(dst, tcg_env, src); |
| gen_store_fpr_F(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FqTOs, ALL, do_env_fq, a, gen_helper_fqtos) |
| TRANS(FqTOi, ALL, do_env_fq, a, gen_helper_fqtoi) |
| |
| static bool do_env_dq(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i64, TCGv_env, TCGv_i128)) |
| { |
| TCGv_i128 src; |
| TCGv_i64 dst; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| src = gen_load_fpr_Q(dc, a->rs); |
| dst = tcg_temp_new_i64(); |
| func(dst, tcg_env, src); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FqTOd, ALL, do_env_dq, a, gen_helper_fqtod) |
| TRANS(FqTOx, 64, do_env_dq, a, gen_helper_fqtox) |
| |
| static bool do_env_qf(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i128, TCGv_env, TCGv_i32)) |
| { |
| TCGv_i32 src; |
| TCGv_i128 dst; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| src = gen_load_fpr_F(dc, a->rs); |
| dst = tcg_temp_new_i128(); |
| func(dst, tcg_env, src); |
| gen_store_fpr_Q(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FiTOq, ALL, do_env_qf, a, gen_helper_fitoq) |
| TRANS(FsTOq, ALL, do_env_qf, a, gen_helper_fstoq) |
| |
| static bool do_env_qd(DisasContext *dc, arg_r_r *a, |
| void (*func)(TCGv_i128, TCGv_env, TCGv_i64)) |
| { |
| TCGv_i64 src; |
| TCGv_i128 dst; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| src = gen_load_fpr_D(dc, a->rs); |
| dst = tcg_temp_new_i128(); |
| func(dst, tcg_env, src); |
| gen_store_fpr_Q(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FdTOq, ALL, do_env_qd, a, gen_helper_fdtoq) |
| TRANS(FxTOq, 64, do_env_qd, a, gen_helper_fxtoq) |
| |
| static bool do_fff(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv_i32, TCGv_i32, TCGv_i32)) |
| { |
| TCGv_i32 src1, src2; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_F(dc, a->rs2); |
| func(src1, src1, src2); |
| gen_store_fpr_F(dc, a->rd, src1); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FPADD16s, VIS1, do_fff, a, tcg_gen_vec_add16_i32) |
| TRANS(FPADD32s, VIS1, do_fff, a, tcg_gen_add_i32) |
| TRANS(FPSUB16s, VIS1, do_fff, a, tcg_gen_vec_sub16_i32) |
| TRANS(FPSUB32s, VIS1, do_fff, a, tcg_gen_sub_i32) |
| TRANS(FNORs, VIS1, do_fff, a, tcg_gen_nor_i32) |
| TRANS(FANDNOTs, VIS1, do_fff, a, tcg_gen_andc_i32) |
| TRANS(FXORs, VIS1, do_fff, a, tcg_gen_xor_i32) |
| TRANS(FNANDs, VIS1, do_fff, a, tcg_gen_nand_i32) |
| TRANS(FANDs, VIS1, do_fff, a, tcg_gen_and_i32) |
| TRANS(FXNORs, VIS1, do_fff, a, tcg_gen_eqv_i32) |
| TRANS(FORNOTs, VIS1, do_fff, a, tcg_gen_orc_i32) |
| TRANS(FORs, VIS1, do_fff, a, tcg_gen_or_i32) |
| |
| TRANS(FHADDs, VIS3, do_fff, a, gen_op_fhadds) |
| TRANS(FHSUBs, VIS3, do_fff, a, gen_op_fhsubs) |
| TRANS(FNHADDs, VIS3, do_fff, a, gen_op_fnhadds) |
| |
| TRANS(FPADDS16s, VIS3, do_fff, a, gen_op_fpadds16s) |
| TRANS(FPSUBS16s, VIS3, do_fff, a, gen_op_fpsubs16s) |
| TRANS(FPADDS32s, VIS3, do_fff, a, gen_op_fpadds32s) |
| TRANS(FPSUBS32s, VIS3, do_fff, a, gen_op_fpsubs32s) |
| |
| static bool do_env_fff(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv_i32, TCGv_env, TCGv_i32, TCGv_i32)) |
| { |
| TCGv_i32 src1, src2; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_F(dc, a->rs2); |
| func(src1, tcg_env, src1, src2); |
| gen_store_fpr_F(dc, a->rd, src1); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FADDs, ALL, do_env_fff, a, gen_helper_fadds) |
| TRANS(FSUBs, ALL, do_env_fff, a, gen_helper_fsubs) |
| TRANS(FMULs, ALL, do_env_fff, a, gen_helper_fmuls) |
| TRANS(FDIVs, ALL, do_env_fff, a, gen_helper_fdivs) |
| TRANS(FNADDs, VIS3, do_env_fff, a, gen_helper_fnadds) |
| TRANS(FNMULs, VIS3, do_env_fff, a, gen_helper_fnmuls) |
| |
| static bool do_dff(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv_i64, TCGv_i32, TCGv_i32)) |
| { |
| TCGv_i64 dst; |
| TCGv_i32 src1, src2; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_F(dc, a->rs2); |
| func(dst, src1, src2); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMUL8x16AU, VIS1, do_dff, a, gen_op_fmul8x16au) |
| TRANS(FMUL8x16AL, VIS1, do_dff, a, gen_op_fmul8x16al) |
| TRANS(FMULD8SUx16, VIS1, do_dff, a, gen_op_fmuld8sux16) |
| TRANS(FMULD8ULx16, VIS1, do_dff, a, gen_op_fmuld8ulx16) |
| TRANS(FPMERGE, VIS1, do_dff, a, gen_helper_fpmerge) |
| |
| static bool do_dfd(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv_i64, TCGv_i32, TCGv_i64)) |
| { |
| TCGv_i64 dst, src2; |
| TCGv_i32 src1; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| func(dst, src1, src2); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMUL8x16, VIS1, do_dfd, a, gen_helper_fmul8x16) |
| |
| static bool do_gvec_ddd(DisasContext *dc, arg_r_r_r *a, MemOp vece, |
| void (*func)(unsigned, uint32_t, uint32_t, |
| uint32_t, uint32_t, uint32_t)) |
| { |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| func(vece, gen_offset_fpr_D(a->rd), gen_offset_fpr_D(a->rs1), |
| gen_offset_fpr_D(a->rs2), 8, 8); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FPADD8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_add) |
| TRANS(FPADD16, VIS1, do_gvec_ddd, a, MO_16, tcg_gen_gvec_add) |
| TRANS(FPADD32, VIS1, do_gvec_ddd, a, MO_32, tcg_gen_gvec_add) |
| |
| TRANS(FPSUB8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_sub) |
| TRANS(FPSUB16, VIS1, do_gvec_ddd, a, MO_16, tcg_gen_gvec_sub) |
| TRANS(FPSUB32, VIS1, do_gvec_ddd, a, MO_32, tcg_gen_gvec_sub) |
| |
| TRANS(FCHKSM16, VIS3, do_gvec_ddd, a, MO_16, gen_op_fchksm16) |
| TRANS(FMEAN16, VIS3, do_gvec_ddd, a, MO_16, gen_op_fmean16) |
| |
| TRANS(FPADDS8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_ssadd) |
| TRANS(FPADDS16, VIS3, do_gvec_ddd, a, MO_16, tcg_gen_gvec_ssadd) |
| TRANS(FPADDS32, VIS3, do_gvec_ddd, a, MO_32, tcg_gen_gvec_ssadd) |
| TRANS(FPADDUS8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_usadd) |
| TRANS(FPADDUS16, VIS4, do_gvec_ddd, a, MO_16, tcg_gen_gvec_usadd) |
| |
| TRANS(FPSUBS8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_sssub) |
| TRANS(FPSUBS16, VIS3, do_gvec_ddd, a, MO_16, tcg_gen_gvec_sssub) |
| TRANS(FPSUBS32, VIS3, do_gvec_ddd, a, MO_32, tcg_gen_gvec_sssub) |
| TRANS(FPSUBUS8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_ussub) |
| TRANS(FPSUBUS16, VIS4, do_gvec_ddd, a, MO_16, tcg_gen_gvec_ussub) |
| |
| TRANS(FSLL16, VIS3, do_gvec_ddd, a, MO_16, tcg_gen_gvec_shlv) |
| TRANS(FSLL32, VIS3, do_gvec_ddd, a, MO_32, tcg_gen_gvec_shlv) |
| TRANS(FSRL16, VIS3, do_gvec_ddd, a, MO_16, tcg_gen_gvec_shrv) |
| TRANS(FSRL32, VIS3, do_gvec_ddd, a, MO_32, tcg_gen_gvec_shrv) |
| TRANS(FSRA16, VIS3, do_gvec_ddd, a, MO_16, tcg_gen_gvec_sarv) |
| TRANS(FSRA32, VIS3, do_gvec_ddd, a, MO_32, tcg_gen_gvec_sarv) |
| |
| TRANS(FPMIN8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_smin) |
| TRANS(FPMIN16, VIS4, do_gvec_ddd, a, MO_16, tcg_gen_gvec_smin) |
| TRANS(FPMIN32, VIS4, do_gvec_ddd, a, MO_32, tcg_gen_gvec_smin) |
| TRANS(FPMINU8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_umin) |
| TRANS(FPMINU16, VIS4, do_gvec_ddd, a, MO_16, tcg_gen_gvec_umin) |
| TRANS(FPMINU32, VIS4, do_gvec_ddd, a, MO_32, tcg_gen_gvec_umin) |
| |
| TRANS(FPMAX8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_smax) |
| TRANS(FPMAX16, VIS4, do_gvec_ddd, a, MO_16, tcg_gen_gvec_smax) |
| TRANS(FPMAX32, VIS4, do_gvec_ddd, a, MO_32, tcg_gen_gvec_smax) |
| TRANS(FPMAXU8, VIS4, do_gvec_ddd, a, MO_8, tcg_gen_gvec_umax) |
| TRANS(FPMAXU16, VIS4, do_gvec_ddd, a, MO_16, tcg_gen_gvec_umax) |
| TRANS(FPMAXU32, VIS4, do_gvec_ddd, a, MO_32, tcg_gen_gvec_umax) |
| |
| static bool do_ddd(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv_i64, TCGv_i64, TCGv_i64)) |
| { |
| TCGv_i64 dst, src1, src2; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src1 = gen_load_fpr_D(dc, a->rs1); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| func(dst, src1, src2); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMUL8SUx16, VIS1, do_ddd, a, gen_helper_fmul8sux16) |
| TRANS(FMUL8ULx16, VIS1, do_ddd, a, gen_helper_fmul8ulx16) |
| |
| TRANS(FNORd, VIS1, do_ddd, a, tcg_gen_nor_i64) |
| TRANS(FANDNOTd, VIS1, do_ddd, a, tcg_gen_andc_i64) |
| TRANS(FXORd, VIS1, do_ddd, a, tcg_gen_xor_i64) |
| TRANS(FNANDd, VIS1, do_ddd, a, tcg_gen_nand_i64) |
| TRANS(FANDd, VIS1, do_ddd, a, tcg_gen_and_i64) |
| TRANS(FXNORd, VIS1, do_ddd, a, tcg_gen_eqv_i64) |
| TRANS(FORNOTd, VIS1, do_ddd, a, tcg_gen_orc_i64) |
| TRANS(FORd, VIS1, do_ddd, a, tcg_gen_or_i64) |
| |
| TRANS(FPACK32, VIS1, do_ddd, a, gen_op_fpack32) |
| TRANS(FALIGNDATAg, VIS1, do_ddd, a, gen_op_faligndata_g) |
| TRANS(BSHUFFLE, VIS2, do_ddd, a, gen_op_bshuffle) |
| |
| TRANS(FHADDd, VIS3, do_ddd, a, gen_op_fhaddd) |
| TRANS(FHSUBd, VIS3, do_ddd, a, gen_op_fhsubd) |
| TRANS(FNHADDd, VIS3, do_ddd, a, gen_op_fnhaddd) |
| |
| TRANS(FPADD64, VIS3B, do_ddd, a, tcg_gen_add_i64) |
| TRANS(FPSUB64, VIS3B, do_ddd, a, tcg_gen_sub_i64) |
| TRANS(FSLAS16, VIS3, do_ddd, a, gen_helper_fslas16) |
| TRANS(FSLAS32, VIS3, do_ddd, a, gen_helper_fslas32) |
| |
| static bool do_rdd(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv, TCGv_i64, TCGv_i64)) |
| { |
| TCGv_i64 src1, src2; |
| TCGv dst; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = gen_dest_gpr(dc, a->rd); |
| src1 = gen_load_fpr_D(dc, a->rs1); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| func(dst, src1, src2); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FPCMPLE16, VIS1, do_rdd, a, gen_helper_fcmple16) |
| TRANS(FPCMPNE16, VIS1, do_rdd, a, gen_helper_fcmpne16) |
| TRANS(FPCMPGT16, VIS1, do_rdd, a, gen_helper_fcmpgt16) |
| TRANS(FPCMPEQ16, VIS1, do_rdd, a, gen_helper_fcmpeq16) |
| TRANS(FPCMPULE16, VIS4, do_rdd, a, gen_helper_fcmpule16) |
| TRANS(FPCMPUGT16, VIS4, do_rdd, a, gen_helper_fcmpugt16) |
| |
| TRANS(FPCMPLE32, VIS1, do_rdd, a, gen_helper_fcmple32) |
| TRANS(FPCMPNE32, VIS1, do_rdd, a, gen_helper_fcmpne32) |
| TRANS(FPCMPGT32, VIS1, do_rdd, a, gen_helper_fcmpgt32) |
| TRANS(FPCMPEQ32, VIS1, do_rdd, a, gen_helper_fcmpeq32) |
| TRANS(FPCMPULE32, VIS4, do_rdd, a, gen_helper_fcmpule32) |
| TRANS(FPCMPUGT32, VIS4, do_rdd, a, gen_helper_fcmpugt32) |
| |
| TRANS(FPCMPEQ8, VIS3B, do_rdd, a, gen_helper_fcmpeq8) |
| TRANS(FPCMPNE8, VIS3B, do_rdd, a, gen_helper_fcmpne8) |
| TRANS(FPCMPULE8, VIS3B, do_rdd, a, gen_helper_fcmpule8) |
| TRANS(FPCMPUGT8, VIS3B, do_rdd, a, gen_helper_fcmpugt8) |
| TRANS(FPCMPLE8, VIS4, do_rdd, a, gen_helper_fcmple8) |
| TRANS(FPCMPGT8, VIS4, do_rdd, a, gen_helper_fcmpgt8) |
| |
| TRANS(PDISTN, VIS3, do_rdd, a, gen_op_pdistn) |
| TRANS(XMULX, VIS3, do_rrr, a, gen_helper_xmulx) |
| TRANS(XMULXHI, VIS3, do_rrr, a, gen_helper_xmulxhi) |
| |
| static bool do_env_ddd(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv_i64, TCGv_env, TCGv_i64, TCGv_i64)) |
| { |
| TCGv_i64 dst, src1, src2; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src1 = gen_load_fpr_D(dc, a->rs1); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| func(dst, tcg_env, src1, src2); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FADDd, ALL, do_env_ddd, a, gen_helper_faddd) |
| TRANS(FSUBd, ALL, do_env_ddd, a, gen_helper_fsubd) |
| TRANS(FMULd, ALL, do_env_ddd, a, gen_helper_fmuld) |
| TRANS(FDIVd, ALL, do_env_ddd, a, gen_helper_fdivd) |
| TRANS(FNADDd, VIS3, do_env_ddd, a, gen_helper_fnaddd) |
| TRANS(FNMULd, VIS3, do_env_ddd, a, gen_helper_fnmuld) |
| |
| static bool trans_FsMULd(DisasContext *dc, arg_r_r_r *a) |
| { |
| TCGv_i64 dst; |
| TCGv_i32 src1, src2; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (!(dc->def->features & CPU_FEATURE_FSMULD)) { |
| return raise_unimpfpop(dc); |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_F(dc, a->rs2); |
| gen_helper_fsmuld(dst, tcg_env, src1, src2); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_FNsMULd(DisasContext *dc, arg_r_r_r *a) |
| { |
| TCGv_i64 dst; |
| TCGv_i32 src1, src2; |
| |
| if (!avail_VIS3(dc)) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| dst = tcg_temp_new_i64(); |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_F(dc, a->rs2); |
| gen_helper_fnsmuld(dst, tcg_env, src1, src2); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool do_ffff(DisasContext *dc, arg_r_r_r_r *a, |
| void (*func)(TCGv_i32, TCGv_i32, TCGv_i32, TCGv_i32)) |
| { |
| TCGv_i32 dst, src1, src2, src3; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_F(dc, a->rs2); |
| src3 = gen_load_fpr_F(dc, a->rs3); |
| dst = tcg_temp_new_i32(); |
| func(dst, src1, src2, src3); |
| gen_store_fpr_F(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMADDs, FMAF, do_ffff, a, gen_op_fmadds) |
| TRANS(FMSUBs, FMAF, do_ffff, a, gen_op_fmsubs) |
| TRANS(FNMSUBs, FMAF, do_ffff, a, gen_op_fnmsubs) |
| TRANS(FNMADDs, FMAF, do_ffff, a, gen_op_fnmadds) |
| |
| static bool do_dddd(DisasContext *dc, arg_r_r_r_r *a, |
| void (*func)(TCGv_i64, TCGv_i64, TCGv_i64, TCGv_i64)) |
| { |
| TCGv_i64 dst, src1, src2, src3; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src1 = gen_load_fpr_D(dc, a->rs1); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| src3 = gen_load_fpr_D(dc, a->rs3); |
| func(dst, src1, src2, src3); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(PDIST, VIS1, do_dddd, a, gen_helper_pdist) |
| TRANS(FMADDd, FMAF, do_dddd, a, gen_op_fmaddd) |
| TRANS(FMSUBd, FMAF, do_dddd, a, gen_op_fmsubd) |
| TRANS(FNMSUBd, FMAF, do_dddd, a, gen_op_fnmsubd) |
| TRANS(FNMADDd, FMAF, do_dddd, a, gen_op_fnmaddd) |
| TRANS(FPMADDX, IMA, do_dddd, a, gen_op_fpmaddx) |
| TRANS(FPMADDXHI, IMA, do_dddd, a, gen_op_fpmaddxhi) |
| |
| static bool trans_FALIGNDATAi(DisasContext *dc, arg_r_r_r *a) |
| { |
| TCGv_i64 dst, src1, src2; |
| TCGv src3; |
| |
| if (!avail_VIS4(dc)) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| dst = tcg_temp_new_i64(); |
| src1 = gen_load_fpr_D(dc, a->rd); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| src3 = gen_load_gpr(dc, a->rs1); |
| gen_op_faligndata_i(dst, src1, src2, src3); |
| gen_store_fpr_D(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool do_env_qqq(DisasContext *dc, arg_r_r_r *a, |
| void (*func)(TCGv_i128, TCGv_env, TCGv_i128, TCGv_i128)) |
| { |
| TCGv_i128 src1, src2; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_Q(dc, a->rs1); |
| src2 = gen_load_fpr_Q(dc, a->rs2); |
| func(src1, tcg_env, src1, src2); |
| gen_store_fpr_Q(dc, a->rd, src1); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FADDq, ALL, do_env_qqq, a, gen_helper_faddq) |
| TRANS(FSUBq, ALL, do_env_qqq, a, gen_helper_fsubq) |
| TRANS(FMULq, ALL, do_env_qqq, a, gen_helper_fmulq) |
| TRANS(FDIVq, ALL, do_env_qqq, a, gen_helper_fdivq) |
| |
| static bool trans_FdMULq(DisasContext *dc, arg_r_r_r *a) |
| { |
| TCGv_i64 src1, src2; |
| TCGv_i128 dst; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_D(dc, a->rs1); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| dst = tcg_temp_new_i128(); |
| gen_helper_fdmulq(dst, tcg_env, src1, src2); |
| gen_store_fpr_Q(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| static bool do_fmovr(DisasContext *dc, arg_FMOVRs *a, bool is_128, |
| void (*func)(DisasContext *, DisasCompare *, int, int)) |
| { |
| DisasCompare cmp; |
| |
| if (!gen_compare_reg(&cmp, a->cond, gen_load_gpr(dc, a->rs1))) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (is_128 && gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| gen_op_clear_ieee_excp_and_FTT(); |
| func(dc, &cmp, a->rd, a->rs2); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMOVRs, 64, do_fmovr, a, false, gen_fmovs) |
| TRANS(FMOVRd, 64, do_fmovr, a, false, gen_fmovd) |
| TRANS(FMOVRq, 64, do_fmovr, a, true, gen_fmovq) |
| |
| static bool do_fmovcc(DisasContext *dc, arg_FMOVscc *a, bool is_128, |
| void (*func)(DisasContext *, DisasCompare *, int, int)) |
| { |
| DisasCompare cmp; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (is_128 && gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| gen_op_clear_ieee_excp_and_FTT(); |
| gen_compare(&cmp, a->cc, a->cond, dc); |
| func(dc, &cmp, a->rd, a->rs2); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMOVscc, 64, do_fmovcc, a, false, gen_fmovs) |
| TRANS(FMOVdcc, 64, do_fmovcc, a, false, gen_fmovd) |
| TRANS(FMOVqcc, 64, do_fmovcc, a, true, gen_fmovq) |
| |
| static bool do_fmovfcc(DisasContext *dc, arg_FMOVsfcc *a, bool is_128, |
| void (*func)(DisasContext *, DisasCompare *, int, int)) |
| { |
| DisasCompare cmp; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (is_128 && gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| gen_op_clear_ieee_excp_and_FTT(); |
| gen_fcompare(&cmp, a->cc, a->cond); |
| func(dc, &cmp, a->rd, a->rs2); |
| return advance_pc(dc); |
| } |
| |
| TRANS(FMOVsfcc, 64, do_fmovfcc, a, false, gen_fmovs) |
| TRANS(FMOVdfcc, 64, do_fmovfcc, a, false, gen_fmovd) |
| TRANS(FMOVqfcc, 64, do_fmovfcc, a, true, gen_fmovq) |
| |
| static bool do_fcmps(DisasContext *dc, arg_FCMPs *a, bool e) |
| { |
| TCGv_i32 src1, src2; |
| |
| if (avail_32(dc) && a->cc != 0) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_F(dc, a->rs2); |
| if (e) { |
| gen_helper_fcmpes(cpu_fcc[a->cc], tcg_env, src1, src2); |
| } else { |
| gen_helper_fcmps(cpu_fcc[a->cc], tcg_env, src1, src2); |
| } |
| return advance_pc(dc); |
| } |
| |
| TRANS(FCMPs, ALL, do_fcmps, a, false) |
| TRANS(FCMPEs, ALL, do_fcmps, a, true) |
| |
| static bool do_fcmpd(DisasContext *dc, arg_FCMPd *a, bool e) |
| { |
| TCGv_i64 src1, src2; |
| |
| if (avail_32(dc) && a->cc != 0) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_D(dc, a->rs1); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| if (e) { |
| gen_helper_fcmped(cpu_fcc[a->cc], tcg_env, src1, src2); |
| } else { |
| gen_helper_fcmpd(cpu_fcc[a->cc], tcg_env, src1, src2); |
| } |
| return advance_pc(dc); |
| } |
| |
| TRANS(FCMPd, ALL, do_fcmpd, a, false) |
| TRANS(FCMPEd, ALL, do_fcmpd, a, true) |
| |
| static bool do_fcmpq(DisasContext *dc, arg_FCMPq *a, bool e) |
| { |
| TCGv_i128 src1, src2; |
| |
| if (avail_32(dc) && a->cc != 0) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| if (gen_trap_float128(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_Q(dc, a->rs1); |
| src2 = gen_load_fpr_Q(dc, a->rs2); |
| if (e) { |
| gen_helper_fcmpeq(cpu_fcc[a->cc], tcg_env, src1, src2); |
| } else { |
| gen_helper_fcmpq(cpu_fcc[a->cc], tcg_env, src1, src2); |
| } |
| return advance_pc(dc); |
| } |
| |
| TRANS(FCMPq, ALL, do_fcmpq, a, false) |
| TRANS(FCMPEq, ALL, do_fcmpq, a, true) |
| |
| static bool trans_FLCMPs(DisasContext *dc, arg_FLCMPs *a) |
| { |
| TCGv_i32 src1, src2; |
| |
| if (!avail_VIS3(dc)) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_F(dc, a->rs1); |
| src2 = gen_load_fpr_F(dc, a->rs2); |
| gen_helper_flcmps(cpu_fcc[a->cc], src1, src2); |
| return advance_pc(dc); |
| } |
| |
| static bool trans_FLCMPd(DisasContext *dc, arg_FLCMPd *a) |
| { |
| TCGv_i64 src1, src2; |
| |
| if (!avail_VIS3(dc)) { |
| return false; |
| } |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| |
| src1 = gen_load_fpr_D(dc, a->rs1); |
| src2 = gen_load_fpr_D(dc, a->rs2); |
| gen_helper_flcmpd(cpu_fcc[a->cc], src1, src2); |
| return advance_pc(dc); |
| } |
| |
| static bool do_movf2r(DisasContext *dc, arg_r_r *a, |
| int (*offset)(unsigned int), |
| void (*load)(TCGv, TCGv_ptr, tcg_target_long)) |
| { |
| TCGv dst; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| dst = gen_dest_gpr(dc, a->rd); |
| load(dst, tcg_env, offset(a->rs)); |
| gen_store_gpr(dc, a->rd, dst); |
| return advance_pc(dc); |
| } |
| |
| TRANS(MOVsTOsw, VIS3B, do_movf2r, a, gen_offset_fpr_F, tcg_gen_ld32s_tl) |
| TRANS(MOVsTOuw, VIS3B, do_movf2r, a, gen_offset_fpr_F, tcg_gen_ld32u_tl) |
| TRANS(MOVdTOx, VIS3B, do_movf2r, a, gen_offset_fpr_D, tcg_gen_ld_tl) |
| |
| static bool do_movr2f(DisasContext *dc, arg_r_r *a, |
| int (*offset)(unsigned int), |
| void (*store)(TCGv, TCGv_ptr, tcg_target_long)) |
| { |
| TCGv src; |
| |
| if (gen_trap_ifnofpu(dc)) { |
| return true; |
| } |
| src = gen_load_gpr(dc, a->rs); |
| store(src, tcg_env, offset(a->rd)); |
| return advance_pc(dc); |
| } |
| |
| TRANS(MOVwTOs, VIS3B, do_movr2f, a, gen_offset_fpr_F, tcg_gen_st32_tl) |
| TRANS(MOVxTOd, VIS3B, do_movr2f, a, gen_offset_fpr_D, tcg_gen_st_tl) |
| |
| static void sparc_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs) |
| { |
| DisasContext *dc = container_of(dcbase, DisasContext, base); |
| int bound; |
| |
| dc->pc = dc->base.pc_first; |
| dc->npc = (target_ulong)dc->base.tb->cs_base; |
| dc->mem_idx = dc->base.tb->flags & TB_FLAG_MMU_MASK; |
| dc->def = &cpu_env(cs)->def; |
| dc->fpu_enabled = tb_fpu_enabled(dc->base.tb->flags); |
| dc->address_mask_32bit = tb_am_enabled(dc->base.tb->flags); |
| #ifndef CONFIG_USER_ONLY |
| dc->supervisor = (dc->base.tb->flags & TB_FLAG_SUPER) != 0; |
| #endif |
| #ifdef TARGET_SPARC64 |
| dc->fprs_dirty = 0; |
| dc->asi = (dc->base.tb->flags >> TB_FLAG_ASI_SHIFT) & 0xff; |
| #ifndef CONFIG_USER_ONLY |
| dc->hypervisor = (dc->base.tb->flags & TB_FLAG_HYPER) != 0; |
| #endif |
| #endif |
| /* |
| * if we reach a page boundary, we stop generation so that the |
| * PC of a TT_TFAULT exception is always in the right page |
| */ |
| bound = -(dc->base.pc_first | TARGET_PAGE_MASK) / 4; |
| dc->base.max_insns = MIN(dc->base.max_insns, bound); |
| } |
| |
| static void sparc_tr_tb_start(DisasContextBase *db, CPUState *cs) |
| { |
| } |
| |
| static void sparc_tr_insn_start(DisasContextBase *dcbase, CPUState *cs) |
| { |
| DisasContext *dc = container_of(dcbase, DisasContext, base); |
| target_ulong npc = dc->npc; |
| |
| if (npc & 3) { |
| switch (npc) { |
| case JUMP_PC: |
| assert(dc->jump_pc[1] == dc->pc + 4); |
| npc = dc->jump_pc[0] | JUMP_PC; |
| break; |
| case DYNAMIC_PC: |
| case DYNAMIC_PC_LOOKUP: |
| npc = DYNAMIC_PC; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| tcg_gen_insn_start(dc->pc, npc); |
| } |
| |
| static void sparc_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs) |
| { |
| DisasContext *dc = container_of(dcbase, DisasContext, base); |
| unsigned int insn; |
| |
| insn = translator_ldl(cpu_env(cs), &dc->base, dc->pc); |
| dc->base.pc_next += 4; |
| |
| if (!decode(dc, insn)) { |
| gen_exception(dc, TT_ILL_INSN); |
| } |
| |
| if (dc->base.is_jmp == DISAS_NORETURN) { |
| return; |
| } |
| if (dc->pc != dc->base.pc_next) { |
| dc->base.is_jmp = DISAS_TOO_MANY; |
| } |
| } |
| |
| static void sparc_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs) |
| { |
| DisasContext *dc = container_of(dcbase, DisasContext, base); |
| DisasDelayException *e, *e_next; |
| bool may_lookup; |
| |
| finishing_insn(dc); |
| |
| switch (dc->base.is_jmp) { |
| case DISAS_NEXT: |
| case DISAS_TOO_MANY: |
| if (((dc->pc | dc->npc) & 3) == 0) { |
| /* static PC and NPC: we can use direct chaining */ |
| gen_goto_tb(dc, 0, dc->pc, dc->npc); |
| break; |
| } |
| |
| may_lookup = true; |
| if (dc->pc & 3) { |
| switch (dc->pc) { |
| case DYNAMIC_PC_LOOKUP: |
| break; |
| case DYNAMIC_PC: |
| may_lookup = false; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } else { |
| tcg_gen_movi_tl(cpu_pc, dc->pc); |
| } |
| |
| if (dc->npc & 3) { |
| switch (dc->npc) { |
| case JUMP_PC: |
| gen_generic_branch(dc); |
| break; |
| case DYNAMIC_PC: |
| may_lookup = false; |
| break; |
| case DYNAMIC_PC_LOOKUP: |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } else { |
| tcg_gen_movi_tl(cpu_npc, dc->npc); |
| } |
| if (may_lookup) { |
| tcg_gen_lookup_and_goto_ptr(); |
| } else { |
| tcg_gen_exit_tb(NULL, 0); |
| } |
| break; |
| |
| case DISAS_NORETURN: |
| break; |
| |
| case DISAS_EXIT: |
| /* Exit TB */ |
| save_state(dc); |
| tcg_gen_exit_tb(NULL, 0); |
| break; |
| |
| default: |
| g_assert_not_reached(); |
| } |
| |
| for (e = dc->delay_excp_list; e ; e = e_next) { |
| gen_set_label(e->lab); |
| |
| tcg_gen_movi_tl(cpu_pc, e->pc); |
| if (e->npc % 4 == 0) { |
| tcg_gen_movi_tl(cpu_npc, e->npc); |
| } |
| gen_helper_raise_exception(tcg_env, e->excp); |
| |
| e_next = e->next; |
| g_free(e); |
| } |
| } |
| |
| static const TranslatorOps sparc_tr_ops = { |
| .init_disas_context = sparc_tr_init_disas_context, |
| .tb_start = sparc_tr_tb_start, |
| .insn_start = sparc_tr_insn_start, |
| .translate_insn = sparc_tr_translate_insn, |
| .tb_stop = sparc_tr_tb_stop, |
| }; |
| |
| void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int *max_insns, |
| vaddr pc, void *host_pc) |
| { |
| DisasContext dc = {}; |
| |
| translator_loop(cs, tb, max_insns, pc, host_pc, &sparc_tr_ops, &dc.base); |
| } |
| |
| void sparc_tcg_init(void) |
| { |
| static const char gregnames[32][4] = { |
| "g0", "g1", "g2", "g3", "g4", "g5", "g6", "g7", |
| "o0", "o1", "o2", "o3", "o4", "o5", "o6", "o7", |
| "l0", "l1", "l2", "l3", "l4", "l5", "l6", "l7", |
| "i0", "i1", "i2", "i3", "i4", "i5", "i6", "i7", |
| }; |
| |
| static const struct { TCGv_i32 *ptr; int off; const char *name; } r32[] = { |
| #ifdef TARGET_SPARC64 |
| { &cpu_fprs, offsetof(CPUSPARCState, fprs), "fprs" }, |
| { &cpu_fcc[0], offsetof(CPUSPARCState, fcc[0]), "fcc0" }, |
| { &cpu_fcc[1], offsetof(CPUSPARCState, fcc[1]), "fcc1" }, |
| { &cpu_fcc[2], offsetof(CPUSPARCState, fcc[2]), "fcc2" }, |
| { &cpu_fcc[3], offsetof(CPUSPARCState, fcc[3]), "fcc3" }, |
| #else |
| { &cpu_fcc[0], offsetof(CPUSPARCState, fcc[0]), "fcc" }, |
| #endif |
| }; |
| |
| static const struct { TCGv *ptr; int off; const char *name; } rtl[] = { |
| #ifdef TARGET_SPARC64 |
| { &cpu_gsr, offsetof(CPUSPARCState, gsr), "gsr" }, |
| { &cpu_xcc_Z, offsetof(CPUSPARCState, xcc_Z), "xcc_Z" }, |
| { &cpu_xcc_C, offsetof(CPUSPARCState, xcc_C), "xcc_C" }, |
| #endif |
| { &cpu_cc_N, offsetof(CPUSPARCState, cc_N), "cc_N" }, |
| { &cpu_cc_V, offsetof(CPUSPARCState, cc_V), "cc_V" }, |
| { &cpu_icc_Z, offsetof(CPUSPARCState, icc_Z), "icc_Z" }, |
| { &cpu_icc_C, offsetof(CPUSPARCState, icc_C), "icc_C" }, |
| { &cpu_cond, offsetof(CPUSPARCState, cond), "cond" }, |
| { &cpu_pc, offsetof(CPUSPARCState, pc), "pc" }, |
| { &cpu_npc, offsetof(CPUSPARCState, npc), "npc" }, |
| { &cpu_y, offsetof(CPUSPARCState, y), "y" }, |
| { &cpu_tbr, offsetof(CPUSPARCState, tbr), "tbr" }, |
| }; |
| |
| unsigned int i; |
| |
| cpu_regwptr = tcg_global_mem_new_ptr(tcg_env, |
| offsetof(CPUSPARCState, regwptr), |
| "regwptr"); |
| |
| for (i = 0; i < ARRAY_SIZE(r32); ++i) { |
| *r32[i].ptr = tcg_global_mem_new_i32(tcg_env, r32[i].off, r32[i].name); |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(rtl); ++i) { |
| *rtl[i].ptr = tcg_global_mem_new(tcg_env, rtl[i].off, rtl[i].name); |
| } |
| |
| cpu_regs[0] = NULL; |
| for (i = 1; i < 8; ++i) { |
| cpu_regs[i] = tcg_global_mem_new(tcg_env, |
| offsetof(CPUSPARCState, gregs[i]), |
| gregnames[i]); |
| } |
| |
| for (i = 8; i < 32; ++i) { |
| cpu_regs[i] = tcg_global_mem_new(cpu_regwptr, |
| (i - 8) * sizeof(target_ulong), |
| gregnames[i]); |
| } |
| } |
| |
| void sparc_restore_state_to_opc(CPUState *cs, |
| const TranslationBlock *tb, |
| const uint64_t *data) |
| { |
| CPUSPARCState *env = cpu_env(cs); |
| target_ulong pc = data[0]; |
| target_ulong npc = data[1]; |
| |
| env->pc = pc; |
| if (npc == DYNAMIC_PC) { |
| /* dynamic NPC: already stored */ |
| } else if (npc & JUMP_PC) { |
| /* jump PC: use 'cond' and the jump targets of the translation */ |
| if (env->cond) { |
| env->npc = npc & ~3; |
| } else { |
| env->npc = pc + 4; |
| } |
| } else { |
| env->npc = npc; |
| } |
| } |