| /* |
| * Alpha emulation cpu translation for qemu. |
| * |
| * Copyright (c) 2007 Jocelyn Mayer |
| * |
| * This library is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2 of the License, or (at your option) any later version. |
| * |
| * This library is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with this library; if not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #include "cpu.h" |
| #include "disas/disas.h" |
| #include "qemu/host-utils.h" |
| #include "tcg-op.h" |
| #include "exec/cpu_ldst.h" |
| |
| #include "exec/helper-proto.h" |
| #include "exec/helper-gen.h" |
| |
| #include "trace-tcg.h" |
| |
| |
| #undef ALPHA_DEBUG_DISAS |
| #define CONFIG_SOFTFLOAT_INLINE |
| |
| #ifdef ALPHA_DEBUG_DISAS |
| # define LOG_DISAS(...) qemu_log_mask(CPU_LOG_TB_IN_ASM, ## __VA_ARGS__) |
| #else |
| # define LOG_DISAS(...) do { } while (0) |
| #endif |
| |
| typedef struct DisasContext DisasContext; |
| struct DisasContext { |
| struct TranslationBlock *tb; |
| uint64_t pc; |
| int mem_idx; |
| |
| /* Current rounding mode for this TB. */ |
| int tb_rm; |
| /* Current flush-to-zero setting for this TB. */ |
| int tb_ftz; |
| |
| /* implver value for this CPU. */ |
| int implver; |
| |
| /* Temporaries for $31 and $f31 as source and destination. */ |
| TCGv zero; |
| TCGv sink; |
| /* Temporary for immediate constants. */ |
| TCGv lit; |
| |
| bool singlestep_enabled; |
| }; |
| |
| /* Return values from translate_one, indicating the state of the TB. |
| Note that zero indicates that we are not exiting the TB. */ |
| |
| typedef enum { |
| NO_EXIT, |
| |
| /* We have emitted one or more goto_tb. No fixup required. */ |
| EXIT_GOTO_TB, |
| |
| /* We are not using a goto_tb (for whatever reason), but have updated |
| the PC (for whatever reason), so there's no need to do it again on |
| exiting the TB. */ |
| EXIT_PC_UPDATED, |
| |
| /* We are exiting the TB, but have neither emitted a goto_tb, nor |
| updated the PC for the next instruction to be executed. */ |
| EXIT_PC_STALE, |
| |
| /* We are ending the TB with a noreturn function call, e.g. longjmp. |
| No following code will be executed. */ |
| EXIT_NORETURN, |
| } ExitStatus; |
| |
| /* global register indexes */ |
| static TCGv_ptr cpu_env; |
| static TCGv cpu_ir[31]; |
| static TCGv cpu_fir[31]; |
| static TCGv cpu_pc; |
| static TCGv cpu_lock_addr; |
| static TCGv cpu_lock_st_addr; |
| static TCGv cpu_lock_value; |
| |
| #include "exec/gen-icount.h" |
| |
| void alpha_translate_init(void) |
| { |
| #define DEF_VAR(V) { &cpu_##V, #V, offsetof(CPUAlphaState, V) } |
| |
| typedef struct { TCGv *var; const char *name; int ofs; } GlobalVar; |
| static const GlobalVar vars[] = { |
| DEF_VAR(pc), |
| DEF_VAR(lock_addr), |
| DEF_VAR(lock_st_addr), |
| DEF_VAR(lock_value), |
| }; |
| |
| #undef DEF_VAR |
| |
| /* Use the symbolic register names that match the disassembler. */ |
| static const char greg_names[31][4] = { |
| "v0", "t0", "t1", "t2", "t3", "t4", "t5", "t6", |
| "t7", "s0", "s1", "s2", "s3", "s4", "s5", "fp", |
| "a0", "a1", "a2", "a3", "a4", "a5", "t8", "t9", |
| "t10", "t11", "ra", "t12", "at", "gp", "sp" |
| }; |
| static const char freg_names[31][4] = { |
| "f0", "f1", "f2", "f3", "f4", "f5", "f6", "f7", |
| "f8", "f9", "f10", "f11", "f12", "f13", "f14", "f15", |
| "f16", "f17", "f18", "f19", "f20", "f21", "f22", "f23", |
| "f24", "f25", "f26", "f27", "f28", "f29", "f30" |
| }; |
| |
| static bool done_init = 0; |
| int i; |
| |
| if (done_init) { |
| return; |
| } |
| done_init = 1; |
| |
| cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env"); |
| |
| for (i = 0; i < 31; i++) { |
| cpu_ir[i] = tcg_global_mem_new_i64(TCG_AREG0, |
| offsetof(CPUAlphaState, ir[i]), |
| greg_names[i]); |
| } |
| |
| for (i = 0; i < 31; i++) { |
| cpu_fir[i] = tcg_global_mem_new_i64(TCG_AREG0, |
| offsetof(CPUAlphaState, fir[i]), |
| freg_names[i]); |
| } |
| |
| for (i = 0; i < ARRAY_SIZE(vars); ++i) { |
| const GlobalVar *v = &vars[i]; |
| *v->var = tcg_global_mem_new_i64(TCG_AREG0, v->ofs, v->name); |
| } |
| } |
| |
| static TCGv load_zero(DisasContext *ctx) |
| { |
| if (TCGV_IS_UNUSED_I64(ctx->zero)) { |
| ctx->zero = tcg_const_i64(0); |
| } |
| return ctx->zero; |
| } |
| |
| static TCGv dest_sink(DisasContext *ctx) |
| { |
| if (TCGV_IS_UNUSED_I64(ctx->sink)) { |
| ctx->sink = tcg_temp_new(); |
| } |
| return ctx->sink; |
| } |
| |
| static TCGv load_gpr(DisasContext *ctx, unsigned reg) |
| { |
| if (likely(reg < 31)) { |
| return cpu_ir[reg]; |
| } else { |
| return load_zero(ctx); |
| } |
| } |
| |
| static TCGv load_gpr_lit(DisasContext *ctx, unsigned reg, |
| uint8_t lit, bool islit) |
| { |
| if (islit) { |
| ctx->lit = tcg_const_i64(lit); |
| return ctx->lit; |
| } else if (likely(reg < 31)) { |
| return cpu_ir[reg]; |
| } else { |
| return load_zero(ctx); |
| } |
| } |
| |
| static TCGv dest_gpr(DisasContext *ctx, unsigned reg) |
| { |
| if (likely(reg < 31)) { |
| return cpu_ir[reg]; |
| } else { |
| return dest_sink(ctx); |
| } |
| } |
| |
| static TCGv load_fpr(DisasContext *ctx, unsigned reg) |
| { |
| if (likely(reg < 31)) { |
| return cpu_fir[reg]; |
| } else { |
| return load_zero(ctx); |
| } |
| } |
| |
| static TCGv dest_fpr(DisasContext *ctx, unsigned reg) |
| { |
| if (likely(reg < 31)) { |
| return cpu_fir[reg]; |
| } else { |
| return dest_sink(ctx); |
| } |
| } |
| |
| static void gen_excp_1(int exception, int error_code) |
| { |
| TCGv_i32 tmp1, tmp2; |
| |
| tmp1 = tcg_const_i32(exception); |
| tmp2 = tcg_const_i32(error_code); |
| gen_helper_excp(cpu_env, tmp1, tmp2); |
| tcg_temp_free_i32(tmp2); |
| tcg_temp_free_i32(tmp1); |
| } |
| |
| static ExitStatus gen_excp(DisasContext *ctx, int exception, int error_code) |
| { |
| tcg_gen_movi_i64(cpu_pc, ctx->pc); |
| gen_excp_1(exception, error_code); |
| return EXIT_NORETURN; |
| } |
| |
| static inline ExitStatus gen_invalid(DisasContext *ctx) |
| { |
| return gen_excp(ctx, EXCP_OPCDEC, 0); |
| } |
| |
| static inline void gen_qemu_ldf(TCGv t0, TCGv t1, int flags) |
| { |
| TCGv_i32 tmp32 = tcg_temp_new_i32(); |
| tcg_gen_qemu_ld_i32(tmp32, t1, flags, MO_LEUL); |
| gen_helper_memory_to_f(t0, tmp32); |
| tcg_temp_free_i32(tmp32); |
| } |
| |
| static inline void gen_qemu_ldg(TCGv t0, TCGv t1, int flags) |
| { |
| TCGv tmp = tcg_temp_new(); |
| tcg_gen_qemu_ld_i64(tmp, t1, flags, MO_LEQ); |
| gen_helper_memory_to_g(t0, tmp); |
| tcg_temp_free(tmp); |
| } |
| |
| static inline void gen_qemu_lds(TCGv t0, TCGv t1, int flags) |
| { |
| TCGv_i32 tmp32 = tcg_temp_new_i32(); |
| tcg_gen_qemu_ld_i32(tmp32, t1, flags, MO_LEUL); |
| gen_helper_memory_to_s(t0, tmp32); |
| tcg_temp_free_i32(tmp32); |
| } |
| |
| static inline void gen_qemu_ldl_l(TCGv t0, TCGv t1, int flags) |
| { |
| tcg_gen_qemu_ld_i64(t0, t1, flags, MO_LESL); |
| tcg_gen_mov_i64(cpu_lock_addr, t1); |
| tcg_gen_mov_i64(cpu_lock_value, t0); |
| } |
| |
| static inline void gen_qemu_ldq_l(TCGv t0, TCGv t1, int flags) |
| { |
| tcg_gen_qemu_ld_i64(t0, t1, flags, MO_LEQ); |
| tcg_gen_mov_i64(cpu_lock_addr, t1); |
| tcg_gen_mov_i64(cpu_lock_value, t0); |
| } |
| |
| static inline void gen_load_mem(DisasContext *ctx, |
| void (*tcg_gen_qemu_load)(TCGv t0, TCGv t1, |
| int flags), |
| int ra, int rb, int32_t disp16, bool fp, |
| bool clear) |
| { |
| TCGv tmp, addr, va; |
| |
| /* LDQ_U with ra $31 is UNOP. Other various loads are forms of |
| prefetches, which we can treat as nops. No worries about |
| missed exceptions here. */ |
| if (unlikely(ra == 31)) { |
| return; |
| } |
| |
| tmp = tcg_temp_new(); |
| addr = load_gpr(ctx, rb); |
| |
| if (disp16) { |
| tcg_gen_addi_i64(tmp, addr, disp16); |
| addr = tmp; |
| } |
| if (clear) { |
| tcg_gen_andi_i64(tmp, addr, ~0x7); |
| addr = tmp; |
| } |
| |
| va = (fp ? cpu_fir[ra] : cpu_ir[ra]); |
| tcg_gen_qemu_load(va, addr, ctx->mem_idx); |
| |
| tcg_temp_free(tmp); |
| } |
| |
| static inline void gen_qemu_stf(TCGv t0, TCGv t1, int flags) |
| { |
| TCGv_i32 tmp32 = tcg_temp_new_i32(); |
| gen_helper_f_to_memory(tmp32, t0); |
| tcg_gen_qemu_st_i32(tmp32, t1, flags, MO_LEUL); |
| tcg_temp_free_i32(tmp32); |
| } |
| |
| static inline void gen_qemu_stg(TCGv t0, TCGv t1, int flags) |
| { |
| TCGv tmp = tcg_temp_new(); |
| gen_helper_g_to_memory(tmp, t0); |
| tcg_gen_qemu_st_i64(tmp, t1, flags, MO_LEQ); |
| tcg_temp_free(tmp); |
| } |
| |
| static inline void gen_qemu_sts(TCGv t0, TCGv t1, int flags) |
| { |
| TCGv_i32 tmp32 = tcg_temp_new_i32(); |
| gen_helper_s_to_memory(tmp32, t0); |
| tcg_gen_qemu_st_i32(tmp32, t1, flags, MO_LEUL); |
| tcg_temp_free_i32(tmp32); |
| } |
| |
| static inline void gen_store_mem(DisasContext *ctx, |
| void (*tcg_gen_qemu_store)(TCGv t0, TCGv t1, |
| int flags), |
| int ra, int rb, int32_t disp16, bool fp, |
| bool clear) |
| { |
| TCGv tmp, addr, va; |
| |
| tmp = tcg_temp_new(); |
| addr = load_gpr(ctx, rb); |
| |
| if (disp16) { |
| tcg_gen_addi_i64(tmp, addr, disp16); |
| addr = tmp; |
| } |
| if (clear) { |
| tcg_gen_andi_i64(tmp, addr, ~0x7); |
| addr = tmp; |
| } |
| |
| va = (fp ? load_fpr(ctx, ra) : load_gpr(ctx, ra)); |
| tcg_gen_qemu_store(va, addr, ctx->mem_idx); |
| |
| tcg_temp_free(tmp); |
| } |
| |
| static ExitStatus gen_store_conditional(DisasContext *ctx, int ra, int rb, |
| int32_t disp16, int quad) |
| { |
| TCGv addr; |
| |
| if (ra == 31) { |
| /* ??? Don't bother storing anything. The user can't tell |
| the difference, since the zero register always reads zero. */ |
| return NO_EXIT; |
| } |
| |
| #if defined(CONFIG_USER_ONLY) |
| addr = cpu_lock_st_addr; |
| #else |
| addr = tcg_temp_local_new(); |
| #endif |
| |
| tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16); |
| |
| #if defined(CONFIG_USER_ONLY) |
| /* ??? This is handled via a complicated version of compare-and-swap |
| in the cpu_loop. Hopefully one day we'll have a real CAS opcode |
| in TCG so that this isn't necessary. */ |
| return gen_excp(ctx, quad ? EXCP_STQ_C : EXCP_STL_C, ra); |
| #else |
| /* ??? In system mode we are never multi-threaded, so CAS can be |
| implemented via a non-atomic load-compare-store sequence. */ |
| { |
| TCGLabel *lab_fail, *lab_done; |
| TCGv val; |
| |
| lab_fail = gen_new_label(); |
| lab_done = gen_new_label(); |
| tcg_gen_brcond_i64(TCG_COND_NE, addr, cpu_lock_addr, lab_fail); |
| |
| val = tcg_temp_new(); |
| tcg_gen_qemu_ld_i64(val, addr, ctx->mem_idx, quad ? MO_LEQ : MO_LESL); |
| tcg_gen_brcond_i64(TCG_COND_NE, val, cpu_lock_value, lab_fail); |
| |
| tcg_gen_qemu_st_i64(cpu_ir[ra], addr, ctx->mem_idx, |
| quad ? MO_LEQ : MO_LEUL); |
| tcg_gen_movi_i64(cpu_ir[ra], 1); |
| tcg_gen_br(lab_done); |
| |
| gen_set_label(lab_fail); |
| tcg_gen_movi_i64(cpu_ir[ra], 0); |
| |
| gen_set_label(lab_done); |
| tcg_gen_movi_i64(cpu_lock_addr, -1); |
| |
| tcg_temp_free(addr); |
| return NO_EXIT; |
| } |
| #endif |
| } |
| |
| static bool in_superpage(DisasContext *ctx, int64_t addr) |
| { |
| return ((ctx->tb->flags & TB_FLAGS_USER_MODE) == 0 |
| && addr < 0 |
| && ((addr >> 41) & 3) == 2 |
| && addr >> TARGET_VIRT_ADDR_SPACE_BITS == addr >> 63); |
| } |
| |
| static bool use_goto_tb(DisasContext *ctx, uint64_t dest) |
| { |
| /* Suppress goto_tb in the case of single-steping and IO. */ |
| if ((ctx->tb->cflags & CF_LAST_IO) |
| || ctx->singlestep_enabled || singlestep) { |
| return false; |
| } |
| /* If the destination is in the superpage, the page perms can't change. */ |
| if (in_superpage(ctx, dest)) { |
| return true; |
| } |
| /* Check for the dest on the same page as the start of the TB. */ |
| return ((ctx->tb->pc ^ dest) & TARGET_PAGE_MASK) == 0; |
| } |
| |
| static ExitStatus gen_bdirect(DisasContext *ctx, int ra, int32_t disp) |
| { |
| uint64_t dest = ctx->pc + (disp << 2); |
| |
| if (ra != 31) { |
| tcg_gen_movi_i64(cpu_ir[ra], ctx->pc); |
| } |
| |
| /* Notice branch-to-next; used to initialize RA with the PC. */ |
| if (disp == 0) { |
| return 0; |
| } else if (use_goto_tb(ctx, dest)) { |
| tcg_gen_goto_tb(0); |
| tcg_gen_movi_i64(cpu_pc, dest); |
| tcg_gen_exit_tb((uintptr_t)ctx->tb); |
| return EXIT_GOTO_TB; |
| } else { |
| tcg_gen_movi_i64(cpu_pc, dest); |
| return EXIT_PC_UPDATED; |
| } |
| } |
| |
| static ExitStatus gen_bcond_internal(DisasContext *ctx, TCGCond cond, |
| TCGv cmp, int32_t disp) |
| { |
| uint64_t dest = ctx->pc + (disp << 2); |
| TCGLabel *lab_true = gen_new_label(); |
| |
| if (use_goto_tb(ctx, dest)) { |
| tcg_gen_brcondi_i64(cond, cmp, 0, lab_true); |
| |
| tcg_gen_goto_tb(0); |
| tcg_gen_movi_i64(cpu_pc, ctx->pc); |
| tcg_gen_exit_tb((uintptr_t)ctx->tb); |
| |
| gen_set_label(lab_true); |
| tcg_gen_goto_tb(1); |
| tcg_gen_movi_i64(cpu_pc, dest); |
| tcg_gen_exit_tb((uintptr_t)ctx->tb + 1); |
| |
| return EXIT_GOTO_TB; |
| } else { |
| TCGv_i64 z = tcg_const_i64(0); |
| TCGv_i64 d = tcg_const_i64(dest); |
| TCGv_i64 p = tcg_const_i64(ctx->pc); |
| |
| tcg_gen_movcond_i64(cond, cpu_pc, cmp, z, d, p); |
| |
| tcg_temp_free_i64(z); |
| tcg_temp_free_i64(d); |
| tcg_temp_free_i64(p); |
| return EXIT_PC_UPDATED; |
| } |
| } |
| |
| static ExitStatus gen_bcond(DisasContext *ctx, TCGCond cond, int ra, |
| int32_t disp, int mask) |
| { |
| TCGv cmp_tmp; |
| |
| if (mask) { |
| cmp_tmp = tcg_temp_new(); |
| tcg_gen_andi_i64(cmp_tmp, load_gpr(ctx, ra), 1); |
| } else { |
| cmp_tmp = load_gpr(ctx, ra); |
| } |
| |
| return gen_bcond_internal(ctx, cond, cmp_tmp, disp); |
| } |
| |
| /* Fold -0.0 for comparison with COND. */ |
| |
| static void gen_fold_mzero(TCGCond cond, TCGv dest, TCGv src) |
| { |
| uint64_t mzero = 1ull << 63; |
| |
| switch (cond) { |
| case TCG_COND_LE: |
| case TCG_COND_GT: |
| /* For <= or >, the -0.0 value directly compares the way we want. */ |
| tcg_gen_mov_i64(dest, src); |
| break; |
| |
| case TCG_COND_EQ: |
| case TCG_COND_NE: |
| /* For == or !=, we can simply mask off the sign bit and compare. */ |
| tcg_gen_andi_i64(dest, src, mzero - 1); |
| break; |
| |
| case TCG_COND_GE: |
| case TCG_COND_LT: |
| /* For >= or <, map -0.0 to +0.0 via comparison and mask. */ |
| tcg_gen_setcondi_i64(TCG_COND_NE, dest, src, mzero); |
| tcg_gen_neg_i64(dest, dest); |
| tcg_gen_and_i64(dest, dest, src); |
| break; |
| |
| default: |
| abort(); |
| } |
| } |
| |
| static ExitStatus gen_fbcond(DisasContext *ctx, TCGCond cond, int ra, |
| int32_t disp) |
| { |
| TCGv cmp_tmp = tcg_temp_new(); |
| gen_fold_mzero(cond, cmp_tmp, load_fpr(ctx, ra)); |
| return gen_bcond_internal(ctx, cond, cmp_tmp, disp); |
| } |
| |
| static void gen_fcmov(DisasContext *ctx, TCGCond cond, int ra, int rb, int rc) |
| { |
| TCGv_i64 va, vb, z; |
| |
| z = load_zero(ctx); |
| vb = load_fpr(ctx, rb); |
| va = tcg_temp_new(); |
| gen_fold_mzero(cond, va, load_fpr(ctx, ra)); |
| |
| tcg_gen_movcond_i64(cond, dest_fpr(ctx, rc), va, z, vb, load_fpr(ctx, rc)); |
| |
| tcg_temp_free(va); |
| } |
| |
| #define QUAL_RM_N 0x080 /* Round mode nearest even */ |
| #define QUAL_RM_C 0x000 /* Round mode chopped */ |
| #define QUAL_RM_M 0x040 /* Round mode minus infinity */ |
| #define QUAL_RM_D 0x0c0 /* Round mode dynamic */ |
| #define QUAL_RM_MASK 0x0c0 |
| |
| #define QUAL_U 0x100 /* Underflow enable (fp output) */ |
| #define QUAL_V 0x100 /* Overflow enable (int output) */ |
| #define QUAL_S 0x400 /* Software completion enable */ |
| #define QUAL_I 0x200 /* Inexact detection enable */ |
| |
| static void gen_qual_roundmode(DisasContext *ctx, int fn11) |
| { |
| TCGv_i32 tmp; |
| |
| fn11 &= QUAL_RM_MASK; |
| if (fn11 == ctx->tb_rm) { |
| return; |
| } |
| ctx->tb_rm = fn11; |
| |
| tmp = tcg_temp_new_i32(); |
| switch (fn11) { |
| case QUAL_RM_N: |
| tcg_gen_movi_i32(tmp, float_round_nearest_even); |
| break; |
| case QUAL_RM_C: |
| tcg_gen_movi_i32(tmp, float_round_to_zero); |
| break; |
| case QUAL_RM_M: |
| tcg_gen_movi_i32(tmp, float_round_down); |
| break; |
| case QUAL_RM_D: |
| tcg_gen_ld8u_i32(tmp, cpu_env, |
| offsetof(CPUAlphaState, fpcr_dyn_round)); |
| break; |
| } |
| |
| #if defined(CONFIG_SOFTFLOAT_INLINE) |
| /* ??? The "fpu/softfloat.h" interface is to call set_float_rounding_mode. |
| With CONFIG_SOFTFLOAT that expands to an out-of-line call that just |
| sets the one field. */ |
| tcg_gen_st8_i32(tmp, cpu_env, |
| offsetof(CPUAlphaState, fp_status.float_rounding_mode)); |
| #else |
| gen_helper_setroundmode(tmp); |
| #endif |
| |
| tcg_temp_free_i32(tmp); |
| } |
| |
| static void gen_qual_flushzero(DisasContext *ctx, int fn11) |
| { |
| TCGv_i32 tmp; |
| |
| fn11 &= QUAL_U; |
| if (fn11 == ctx->tb_ftz) { |
| return; |
| } |
| ctx->tb_ftz = fn11; |
| |
| tmp = tcg_temp_new_i32(); |
| if (fn11) { |
| /* Underflow is enabled, use the FPCR setting. */ |
| tcg_gen_ld8u_i32(tmp, cpu_env, |
| offsetof(CPUAlphaState, fpcr_flush_to_zero)); |
| } else { |
| /* Underflow is disabled, force flush-to-zero. */ |
| tcg_gen_movi_i32(tmp, 1); |
| } |
| |
| #if defined(CONFIG_SOFTFLOAT_INLINE) |
| tcg_gen_st8_i32(tmp, cpu_env, |
| offsetof(CPUAlphaState, fp_status.flush_to_zero)); |
| #else |
| gen_helper_setflushzero(tmp); |
| #endif |
| |
| tcg_temp_free_i32(tmp); |
| } |
| |
| static TCGv gen_ieee_input(DisasContext *ctx, int reg, int fn11, int is_cmp) |
| { |
| TCGv val; |
| |
| if (unlikely(reg == 31)) { |
| val = load_zero(ctx); |
| } else { |
| val = cpu_fir[reg]; |
| if ((fn11 & QUAL_S) == 0) { |
| if (is_cmp) { |
| gen_helper_ieee_input_cmp(cpu_env, val); |
| } else { |
| gen_helper_ieee_input(cpu_env, val); |
| } |
| } else { |
| #ifndef CONFIG_USER_ONLY |
| /* In system mode, raise exceptions for denormals like real |
| hardware. In user mode, proceed as if the OS completion |
| handler is handling the denormal as per spec. */ |
| gen_helper_ieee_input_s(cpu_env, val); |
| #endif |
| } |
| } |
| return val; |
| } |
| |
| static void gen_fp_exc_raise(int rc, int fn11) |
| { |
| /* ??? We ought to be able to do something with imprecise exceptions. |
| E.g. notice we're still in the trap shadow of something within the |
| TB and do not generate the code to signal the exception; end the TB |
| when an exception is forced to arrive, either by consumption of a |
| register value or TRAPB or EXCB. */ |
| TCGv_i32 reg, ign; |
| uint32_t ignore = 0; |
| |
| if (!(fn11 & QUAL_U)) { |
| /* Note that QUAL_U == QUAL_V, so ignore either. */ |
| ignore |= FPCR_UNF | FPCR_IOV; |
| } |
| if (!(fn11 & QUAL_I)) { |
| ignore |= FPCR_INE; |
| } |
| ign = tcg_const_i32(ignore); |
| |
| /* ??? Pass in the regno of the destination so that the helper can |
| set EXC_MASK, which contains a bitmask of destination registers |
| that have caused arithmetic traps. A simple userspace emulation |
| does not require this. We do need it for a guest kernel's entArith, |
| or if we were to do something clever with imprecise exceptions. */ |
| reg = tcg_const_i32(rc + 32); |
| if (fn11 & QUAL_S) { |
| gen_helper_fp_exc_raise_s(cpu_env, ign, reg); |
| } else { |
| gen_helper_fp_exc_raise(cpu_env, ign, reg); |
| } |
| |
| tcg_temp_free_i32(reg); |
| tcg_temp_free_i32(ign); |
| } |
| |
| static void gen_cvtlq(TCGv vc, TCGv vb) |
| { |
| TCGv tmp = tcg_temp_new(); |
| |
| /* The arithmetic right shift here, plus the sign-extended mask below |
| yields a sign-extended result without an explicit ext32s_i64. */ |
| tcg_gen_sari_i64(tmp, vb, 32); |
| tcg_gen_shri_i64(vc, vb, 29); |
| tcg_gen_andi_i64(tmp, tmp, (int32_t)0xc0000000); |
| tcg_gen_andi_i64(vc, vc, 0x3fffffff); |
| tcg_gen_or_i64(vc, vc, tmp); |
| |
| tcg_temp_free(tmp); |
| } |
| |
| static void gen_ieee_arith2(DisasContext *ctx, |
| void (*helper)(TCGv, TCGv_ptr, TCGv), |
| int rb, int rc, int fn11) |
| { |
| TCGv vb; |
| |
| gen_qual_roundmode(ctx, fn11); |
| gen_qual_flushzero(ctx, fn11); |
| |
| vb = gen_ieee_input(ctx, rb, fn11, 0); |
| helper(dest_fpr(ctx, rc), cpu_env, vb); |
| |
| gen_fp_exc_raise(rc, fn11); |
| } |
| |
| #define IEEE_ARITH2(name) \ |
| static inline void glue(gen_, name)(DisasContext *ctx, \ |
| int rb, int rc, int fn11) \ |
| { \ |
| gen_ieee_arith2(ctx, gen_helper_##name, rb, rc, fn11); \ |
| } |
| IEEE_ARITH2(sqrts) |
| IEEE_ARITH2(sqrtt) |
| IEEE_ARITH2(cvtst) |
| IEEE_ARITH2(cvtts) |
| |
| static void gen_cvttq(DisasContext *ctx, int rb, int rc, int fn11) |
| { |
| TCGv vb, vc; |
| |
| /* No need to set flushzero, since we have an integer output. */ |
| vb = gen_ieee_input(ctx, rb, fn11, 0); |
| vc = dest_fpr(ctx, rc); |
| |
| /* Almost all integer conversions use cropped rounding; |
| special case that. */ |
| if ((fn11 & QUAL_RM_MASK) == QUAL_RM_C) { |
| gen_helper_cvttq_c(vc, cpu_env, vb); |
| } else { |
| gen_qual_roundmode(ctx, fn11); |
| gen_helper_cvttq(vc, cpu_env, vb); |
| } |
| gen_fp_exc_raise(rc, fn11); |
| } |
| |
| static void gen_ieee_intcvt(DisasContext *ctx, |
| void (*helper)(TCGv, TCGv_ptr, TCGv), |
| int rb, int rc, int fn11) |
| { |
| TCGv vb, vc; |
| |
| gen_qual_roundmode(ctx, fn11); |
| vb = load_fpr(ctx, rb); |
| vc = dest_fpr(ctx, rc); |
| |
| /* The only exception that can be raised by integer conversion |
| is inexact. Thus we only need to worry about exceptions when |
| inexact handling is requested. */ |
| if (fn11 & QUAL_I) { |
| helper(vc, cpu_env, vb); |
| gen_fp_exc_raise(rc, fn11); |
| } else { |
| helper(vc, cpu_env, vb); |
| } |
| } |
| |
| #define IEEE_INTCVT(name) \ |
| static inline void glue(gen_, name)(DisasContext *ctx, \ |
| int rb, int rc, int fn11) \ |
| { \ |
| gen_ieee_intcvt(ctx, gen_helper_##name, rb, rc, fn11); \ |
| } |
| IEEE_INTCVT(cvtqs) |
| IEEE_INTCVT(cvtqt) |
| |
| static void gen_cpy_mask(TCGv vc, TCGv va, TCGv vb, bool inv_a, uint64_t mask) |
| { |
| TCGv vmask = tcg_const_i64(mask); |
| TCGv tmp = tcg_temp_new_i64(); |
| |
| if (inv_a) { |
| tcg_gen_andc_i64(tmp, vmask, va); |
| } else { |
| tcg_gen_and_i64(tmp, va, vmask); |
| } |
| |
| tcg_gen_andc_i64(vc, vb, vmask); |
| tcg_gen_or_i64(vc, vc, tmp); |
| |
| tcg_temp_free(vmask); |
| tcg_temp_free(tmp); |
| } |
| |
| static void gen_ieee_arith3(DisasContext *ctx, |
| void (*helper)(TCGv, TCGv_ptr, TCGv, TCGv), |
| int ra, int rb, int rc, int fn11) |
| { |
| TCGv va, vb, vc; |
| |
| gen_qual_roundmode(ctx, fn11); |
| gen_qual_flushzero(ctx, fn11); |
| |
| va = gen_ieee_input(ctx, ra, fn11, 0); |
| vb = gen_ieee_input(ctx, rb, fn11, 0); |
| vc = dest_fpr(ctx, rc); |
| helper(vc, cpu_env, va, vb); |
| |
| gen_fp_exc_raise(rc, fn11); |
| } |
| |
| #define IEEE_ARITH3(name) \ |
| static inline void glue(gen_, name)(DisasContext *ctx, \ |
| int ra, int rb, int rc, int fn11) \ |
| { \ |
| gen_ieee_arith3(ctx, gen_helper_##name, ra, rb, rc, fn11); \ |
| } |
| IEEE_ARITH3(adds) |
| IEEE_ARITH3(subs) |
| IEEE_ARITH3(muls) |
| IEEE_ARITH3(divs) |
| IEEE_ARITH3(addt) |
| IEEE_ARITH3(subt) |
| IEEE_ARITH3(mult) |
| IEEE_ARITH3(divt) |
| |
| static void gen_ieee_compare(DisasContext *ctx, |
| void (*helper)(TCGv, TCGv_ptr, TCGv, TCGv), |
| int ra, int rb, int rc, int fn11) |
| { |
| TCGv va, vb, vc; |
| |
| va = gen_ieee_input(ctx, ra, fn11, 1); |
| vb = gen_ieee_input(ctx, rb, fn11, 1); |
| vc = dest_fpr(ctx, rc); |
| helper(vc, cpu_env, va, vb); |
| |
| gen_fp_exc_raise(rc, fn11); |
| } |
| |
| #define IEEE_CMP3(name) \ |
| static inline void glue(gen_, name)(DisasContext *ctx, \ |
| int ra, int rb, int rc, int fn11) \ |
| { \ |
| gen_ieee_compare(ctx, gen_helper_##name, ra, rb, rc, fn11); \ |
| } |
| IEEE_CMP3(cmptun) |
| IEEE_CMP3(cmpteq) |
| IEEE_CMP3(cmptlt) |
| IEEE_CMP3(cmptle) |
| |
| static inline uint64_t zapnot_mask(uint8_t lit) |
| { |
| uint64_t mask = 0; |
| int i; |
| |
| for (i = 0; i < 8; ++i) { |
| if ((lit >> i) & 1) { |
| mask |= 0xffull << (i * 8); |
| } |
| } |
| return mask; |
| } |
| |
| /* Implement zapnot with an immediate operand, which expands to some |
| form of immediate AND. This is a basic building block in the |
| definition of many of the other byte manipulation instructions. */ |
| static void gen_zapnoti(TCGv dest, TCGv src, uint8_t lit) |
| { |
| switch (lit) { |
| case 0x00: |
| tcg_gen_movi_i64(dest, 0); |
| break; |
| case 0x01: |
| tcg_gen_ext8u_i64(dest, src); |
| break; |
| case 0x03: |
| tcg_gen_ext16u_i64(dest, src); |
| break; |
| case 0x0f: |
| tcg_gen_ext32u_i64(dest, src); |
| break; |
| case 0xff: |
| tcg_gen_mov_i64(dest, src); |
| break; |
| default: |
| tcg_gen_andi_i64(dest, src, zapnot_mask(lit)); |
| break; |
| } |
| } |
| |
| /* EXTWH, EXTLH, EXTQH */ |
| static void gen_ext_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit, |
| uint8_t lit, uint8_t byte_mask) |
| { |
| if (islit) { |
| tcg_gen_shli_i64(vc, va, (64 - lit * 8) & 0x3f); |
| } else { |
| TCGv tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, load_gpr(ctx, rb), 3); |
| tcg_gen_neg_i64(tmp, tmp); |
| tcg_gen_andi_i64(tmp, tmp, 0x3f); |
| tcg_gen_shl_i64(vc, va, tmp); |
| tcg_temp_free(tmp); |
| } |
| gen_zapnoti(vc, vc, byte_mask); |
| } |
| |
| /* EXTBL, EXTWL, EXTLL, EXTQL */ |
| static void gen_ext_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit, |
| uint8_t lit, uint8_t byte_mask) |
| { |
| if (islit) { |
| tcg_gen_shri_i64(vc, va, (lit & 7) * 8); |
| } else { |
| TCGv tmp = tcg_temp_new(); |
| tcg_gen_andi_i64(tmp, load_gpr(ctx, rb), 7); |
| tcg_gen_shli_i64(tmp, tmp, 3); |
| tcg_gen_shr_i64(vc, va, tmp); |
| tcg_temp_free(tmp); |
| } |
| gen_zapnoti(vc, vc, byte_mask); |
| } |
| |
| /* INSWH, INSLH, INSQH */ |
| static void gen_ins_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit, |
| uint8_t lit, uint8_t byte_mask) |
| { |
| TCGv tmp = tcg_temp_new(); |
| |
| /* The instruction description has us left-shift the byte mask and extract |
| bits <15:8> and apply that zap at the end. This is equivalent to simply |
| performing the zap first and shifting afterward. */ |
| gen_zapnoti(tmp, va, byte_mask); |
| |
| if (islit) { |
| lit &= 7; |
| if (unlikely(lit == 0)) { |
| tcg_gen_movi_i64(vc, 0); |
| } else { |
| tcg_gen_shri_i64(vc, tmp, 64 - lit * 8); |
| } |
| } else { |
| TCGv shift = tcg_temp_new(); |
| |
| /* If (B & 7) == 0, we need to shift by 64 and leave a zero. Do this |
| portably by splitting the shift into two parts: shift_count-1 and 1. |
| Arrange for the -1 by using ones-complement instead of |
| twos-complement in the negation: ~(B * 8) & 63. */ |
| |
| tcg_gen_shli_i64(shift, load_gpr(ctx, rb), 3); |
| tcg_gen_not_i64(shift, shift); |
| tcg_gen_andi_i64(shift, shift, 0x3f); |
| |
| tcg_gen_shr_i64(vc, tmp, shift); |
| tcg_gen_shri_i64(vc, vc, 1); |
| tcg_temp_free(shift); |
| } |
| tcg_temp_free(tmp); |
| } |
| |
| /* INSBL, INSWL, INSLL, INSQL */ |
| static void gen_ins_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit, |
| uint8_t lit, uint8_t byte_mask) |
| { |
| TCGv tmp = tcg_temp_new(); |
| |
| /* The instruction description has us left-shift the byte mask |
| the same number of byte slots as the data and apply the zap |
| at the end. This is equivalent to simply performing the zap |
| first and shifting afterward. */ |
| gen_zapnoti(tmp, va, byte_mask); |
| |
| if (islit) { |
| tcg_gen_shli_i64(vc, tmp, (lit & 7) * 8); |
| } else { |
| TCGv shift = tcg_temp_new(); |
| tcg_gen_andi_i64(shift, load_gpr(ctx, rb), 7); |
| tcg_gen_shli_i64(shift, shift, 3); |
| tcg_gen_shl_i64(vc, tmp, shift); |
| tcg_temp_free(shift); |
| } |
| tcg_temp_free(tmp); |
| } |
| |
| /* MSKWH, MSKLH, MSKQH */ |
| static void gen_msk_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit, |
| uint8_t lit, uint8_t byte_mask) |
| { |
| if (islit) { |
| gen_zapnoti(vc, va, ~((byte_mask << (lit & 7)) >> 8)); |
| } else { |
| TCGv shift = tcg_temp_new(); |
| TCGv mask = tcg_temp_new(); |
| |
| /* The instruction description is as above, where the byte_mask |
| is shifted left, and then we extract bits <15:8>. This can be |
| emulated with a right-shift on the expanded byte mask. This |
| requires extra care because for an input <2:0> == 0 we need a |
| shift of 64 bits in order to generate a zero. This is done by |
| splitting the shift into two parts, the variable shift - 1 |
| followed by a constant 1 shift. The code we expand below is |
| equivalent to ~(B * 8) & 63. */ |
| |
| tcg_gen_shli_i64(shift, load_gpr(ctx, rb), 3); |
| tcg_gen_not_i64(shift, shift); |
| tcg_gen_andi_i64(shift, shift, 0x3f); |
| tcg_gen_movi_i64(mask, zapnot_mask (byte_mask)); |
| tcg_gen_shr_i64(mask, mask, shift); |
| tcg_gen_shri_i64(mask, mask, 1); |
| |
| tcg_gen_andc_i64(vc, va, mask); |
| |
| tcg_temp_free(mask); |
| tcg_temp_free(shift); |
| } |
| } |
| |
| /* MSKBL, MSKWL, MSKLL, MSKQL */ |
| static void gen_msk_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit, |
| uint8_t lit, uint8_t byte_mask) |
| { |
| if (islit) { |
| gen_zapnoti(vc, va, ~(byte_mask << (lit & 7))); |
| } else { |
| TCGv shift = tcg_temp_new(); |
| TCGv mask = tcg_temp_new(); |
| |
| tcg_gen_andi_i64(shift, load_gpr(ctx, rb), 7); |
| tcg_gen_shli_i64(shift, shift, 3); |
| tcg_gen_movi_i64(mask, zapnot_mask(byte_mask)); |
| tcg_gen_shl_i64(mask, mask, shift); |
| |
| tcg_gen_andc_i64(vc, va, mask); |
| |
| tcg_temp_free(mask); |
| tcg_temp_free(shift); |
| } |
| } |
| |
| static void gen_rx(int ra, int set) |
| { |
| TCGv_i32 tmp; |
| |
| if (ra != 31) { |
| tcg_gen_ld8u_i64(cpu_ir[ra], cpu_env, offsetof(CPUAlphaState, intr_flag)); |
| } |
| |
| tmp = tcg_const_i32(set); |
| tcg_gen_st8_i32(tmp, cpu_env, offsetof(CPUAlphaState, intr_flag)); |
| tcg_temp_free_i32(tmp); |
| } |
| |
| static ExitStatus gen_call_pal(DisasContext *ctx, int palcode) |
| { |
| /* We're emulating OSF/1 PALcode. Many of these are trivial access |
| to internal cpu registers. */ |
| |
| /* Unprivileged PAL call */ |
| if (palcode >= 0x80 && palcode < 0xC0) { |
| switch (palcode) { |
| case 0x86: |
| /* IMB */ |
| /* No-op inside QEMU. */ |
| break; |
| case 0x9E: |
| /* RDUNIQUE */ |
| tcg_gen_ld_i64(cpu_ir[IR_V0], cpu_env, |
| offsetof(CPUAlphaState, unique)); |
| break; |
| case 0x9F: |
| /* WRUNIQUE */ |
| tcg_gen_st_i64(cpu_ir[IR_A0], cpu_env, |
| offsetof(CPUAlphaState, unique)); |
| break; |
| default: |
| palcode &= 0xbf; |
| goto do_call_pal; |
| } |
| return NO_EXIT; |
| } |
| |
| #ifndef CONFIG_USER_ONLY |
| /* Privileged PAL code */ |
| if (palcode < 0x40 && (ctx->tb->flags & TB_FLAGS_USER_MODE) == 0) { |
| switch (palcode) { |
| case 0x01: |
| /* CFLUSH */ |
| /* No-op inside QEMU. */ |
| break; |
| case 0x02: |
| /* DRAINA */ |
| /* No-op inside QEMU. */ |
| break; |
| case 0x2D: |
| /* WRVPTPTR */ |
| tcg_gen_st_i64(cpu_ir[IR_A0], cpu_env, |
| offsetof(CPUAlphaState, vptptr)); |
| break; |
| case 0x31: |
| /* WRVAL */ |
| tcg_gen_st_i64(cpu_ir[IR_A0], cpu_env, |
| offsetof(CPUAlphaState, sysval)); |
| break; |
| case 0x32: |
| /* RDVAL */ |
| tcg_gen_ld_i64(cpu_ir[IR_V0], cpu_env, |
| offsetof(CPUAlphaState, sysval)); |
| break; |
| |
| case 0x35: { |
| /* SWPIPL */ |
| TCGv tmp; |
| |
| /* Note that we already know we're in kernel mode, so we know |
| that PS only contains the 3 IPL bits. */ |
| tcg_gen_ld8u_i64(cpu_ir[IR_V0], cpu_env, |
| offsetof(CPUAlphaState, ps)); |
| |
| /* But make sure and store only the 3 IPL bits from the user. */ |
| tmp = tcg_temp_new(); |
| tcg_gen_andi_i64(tmp, cpu_ir[IR_A0], PS_INT_MASK); |
| tcg_gen_st8_i64(tmp, cpu_env, offsetof(CPUAlphaState, ps)); |
| tcg_temp_free(tmp); |
| break; |
| } |
| |
| case 0x36: |
| /* RDPS */ |
| tcg_gen_ld8u_i64(cpu_ir[IR_V0], cpu_env, |
| offsetof(CPUAlphaState, ps)); |
| break; |
| case 0x38: |
| /* WRUSP */ |
| tcg_gen_st_i64(cpu_ir[IR_A0], cpu_env, |
| offsetof(CPUAlphaState, usp)); |
| break; |
| case 0x3A: |
| /* RDUSP */ |
| tcg_gen_ld_i64(cpu_ir[IR_V0], cpu_env, |
| offsetof(CPUAlphaState, usp)); |
| break; |
| case 0x3C: |
| /* WHAMI */ |
| tcg_gen_ld32s_i64(cpu_ir[IR_V0], cpu_env, |
| -offsetof(AlphaCPU, env) + offsetof(CPUState, cpu_index)); |
| break; |
| |
| default: |
| palcode &= 0x3f; |
| goto do_call_pal; |
| } |
| return NO_EXIT; |
| } |
| #endif |
| return gen_invalid(ctx); |
| |
| do_call_pal: |
| #ifdef CONFIG_USER_ONLY |
| return gen_excp(ctx, EXCP_CALL_PAL, palcode); |
| #else |
| { |
| TCGv pc = tcg_const_i64(ctx->pc); |
| TCGv entry = tcg_const_i64(palcode & 0x80 |
| ? 0x2000 + (palcode - 0x80) * 64 |
| : 0x1000 + palcode * 64); |
| |
| gen_helper_call_pal(cpu_env, pc, entry); |
| |
| tcg_temp_free(entry); |
| tcg_temp_free(pc); |
| |
| /* Since the destination is running in PALmode, we don't really |
| need the page permissions check. We'll see the existence of |
| the page when we create the TB, and we'll flush all TBs if |
| we change the PAL base register. */ |
| if (!ctx->singlestep_enabled && !(ctx->tb->cflags & CF_LAST_IO)) { |
| tcg_gen_goto_tb(0); |
| tcg_gen_exit_tb((uintptr_t)ctx->tb); |
| return EXIT_GOTO_TB; |
| } |
| |
| return EXIT_PC_UPDATED; |
| } |
| #endif |
| } |
| |
| #ifndef CONFIG_USER_ONLY |
| |
| #define PR_BYTE 0x100000 |
| #define PR_LONG 0x200000 |
| |
| static int cpu_pr_data(int pr) |
| { |
| switch (pr) { |
| case 0: return offsetof(CPUAlphaState, ps) | PR_BYTE; |
| case 1: return offsetof(CPUAlphaState, fen) | PR_BYTE; |
| case 2: return offsetof(CPUAlphaState, pcc_ofs) | PR_LONG; |
| case 3: return offsetof(CPUAlphaState, trap_arg0); |
| case 4: return offsetof(CPUAlphaState, trap_arg1); |
| case 5: return offsetof(CPUAlphaState, trap_arg2); |
| case 6: return offsetof(CPUAlphaState, exc_addr); |
| case 7: return offsetof(CPUAlphaState, palbr); |
| case 8: return offsetof(CPUAlphaState, ptbr); |
| case 9: return offsetof(CPUAlphaState, vptptr); |
| case 10: return offsetof(CPUAlphaState, unique); |
| case 11: return offsetof(CPUAlphaState, sysval); |
| case 12: return offsetof(CPUAlphaState, usp); |
| |
| case 32 ... 39: |
| return offsetof(CPUAlphaState, shadow[pr - 32]); |
| case 40 ... 63: |
| return offsetof(CPUAlphaState, scratch[pr - 40]); |
| |
| case 251: |
| return offsetof(CPUAlphaState, alarm_expire); |
| } |
| return 0; |
| } |
| |
| static ExitStatus gen_mfpr(DisasContext *ctx, TCGv va, int regno) |
| { |
| int data = cpu_pr_data(regno); |
| |
| /* Special help for VMTIME and WALLTIME. */ |
| if (regno == 250 || regno == 249) { |
| void (*helper)(TCGv) = gen_helper_get_walltime; |
| if (regno == 249) { |
| helper = gen_helper_get_vmtime; |
| } |
| if (ctx->tb->cflags & CF_USE_ICOUNT) { |
| gen_io_start(); |
| helper(va); |
| gen_io_end(); |
| return EXIT_PC_STALE; |
| } else { |
| helper(va); |
| return NO_EXIT; |
| } |
| } |
| |
| /* The basic registers are data only, and unknown registers |
| are read-zero, write-ignore. */ |
| if (data == 0) { |
| tcg_gen_movi_i64(va, 0); |
| } else if (data & PR_BYTE) { |
| tcg_gen_ld8u_i64(va, cpu_env, data & ~PR_BYTE); |
| } else if (data & PR_LONG) { |
| tcg_gen_ld32s_i64(va, cpu_env, data & ~PR_LONG); |
| } else { |
| tcg_gen_ld_i64(va, cpu_env, data); |
| } |
| return NO_EXIT; |
| } |
| |
| static ExitStatus gen_mtpr(DisasContext *ctx, TCGv vb, int regno) |
| { |
| TCGv tmp; |
| int data; |
| |
| switch (regno) { |
| case 255: |
| /* TBIA */ |
| gen_helper_tbia(cpu_env); |
| break; |
| |
| case 254: |
| /* TBIS */ |
| gen_helper_tbis(cpu_env, vb); |
| break; |
| |
| case 253: |
| /* WAIT */ |
| tmp = tcg_const_i64(1); |
| tcg_gen_st32_i64(tmp, cpu_env, -offsetof(AlphaCPU, env) + |
| offsetof(CPUState, halted)); |
| return gen_excp(ctx, EXCP_HLT, 0); |
| |
| case 252: |
| /* HALT */ |
| gen_helper_halt(vb); |
| return EXIT_PC_STALE; |
| |
| case 251: |
| /* ALARM */ |
| gen_helper_set_alarm(cpu_env, vb); |
| break; |
| |
| case 7: |
| /* PALBR */ |
| tcg_gen_st_i64(vb, cpu_env, offsetof(CPUAlphaState, palbr)); |
| /* Changing the PAL base register implies un-chaining all of the TBs |
| that ended with a CALL_PAL. Since the base register usually only |
| changes during boot, flushing everything works well. */ |
| gen_helper_tb_flush(cpu_env); |
| return EXIT_PC_STALE; |
| |
| default: |
| /* The basic registers are data only, and unknown registers |
| are read-zero, write-ignore. */ |
| data = cpu_pr_data(regno); |
| if (data != 0) { |
| if (data & PR_BYTE) { |
| tcg_gen_st8_i64(vb, cpu_env, data & ~PR_BYTE); |
| } else if (data & PR_LONG) { |
| tcg_gen_st32_i64(vb, cpu_env, data & ~PR_LONG); |
| } else { |
| tcg_gen_st_i64(vb, cpu_env, data); |
| } |
| } |
| break; |
| } |
| |
| return NO_EXIT; |
| } |
| #endif /* !USER_ONLY*/ |
| |
| #define REQUIRE_NO_LIT \ |
| do { \ |
| if (real_islit) { \ |
| goto invalid_opc; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_TB_FLAG(FLAG) \ |
| do { \ |
| if ((ctx->tb->flags & (FLAG)) == 0) { \ |
| goto invalid_opc; \ |
| } \ |
| } while (0) |
| |
| #define REQUIRE_REG_31(WHICH) \ |
| do { \ |
| if (WHICH != 31) { \ |
| goto invalid_opc; \ |
| } \ |
| } while (0) |
| |
| static ExitStatus translate_one(DisasContext *ctx, uint32_t insn) |
| { |
| int32_t disp21, disp16, disp12 __attribute__((unused)); |
| uint16_t fn11; |
| uint8_t opc, ra, rb, rc, fpfn, fn7, lit; |
| bool islit, real_islit; |
| TCGv va, vb, vc, tmp, tmp2; |
| TCGv_i32 t32; |
| ExitStatus ret; |
| |
| /* Decode all instruction fields */ |
| opc = extract32(insn, 26, 6); |
| ra = extract32(insn, 21, 5); |
| rb = extract32(insn, 16, 5); |
| rc = extract32(insn, 0, 5); |
| real_islit = islit = extract32(insn, 12, 1); |
| lit = extract32(insn, 13, 8); |
| |
| disp21 = sextract32(insn, 0, 21); |
| disp16 = sextract32(insn, 0, 16); |
| disp12 = sextract32(insn, 0, 12); |
| |
| fn11 = extract32(insn, 5, 11); |
| fpfn = extract32(insn, 5, 6); |
| fn7 = extract32(insn, 5, 7); |
| |
| if (rb == 31 && !islit) { |
| islit = true; |
| lit = 0; |
| } |
| |
| ret = NO_EXIT; |
| switch (opc) { |
| case 0x00: |
| /* CALL_PAL */ |
| ret = gen_call_pal(ctx, insn & 0x03ffffff); |
| break; |
| case 0x01: |
| /* OPC01 */ |
| goto invalid_opc; |
| case 0x02: |
| /* OPC02 */ |
| goto invalid_opc; |
| case 0x03: |
| /* OPC03 */ |
| goto invalid_opc; |
| case 0x04: |
| /* OPC04 */ |
| goto invalid_opc; |
| case 0x05: |
| /* OPC05 */ |
| goto invalid_opc; |
| case 0x06: |
| /* OPC06 */ |
| goto invalid_opc; |
| case 0x07: |
| /* OPC07 */ |
| goto invalid_opc; |
| |
| case 0x09: |
| /* LDAH */ |
| disp16 = (uint32_t)disp16 << 16; |
| /* fall through */ |
| case 0x08: |
| /* LDA */ |
| va = dest_gpr(ctx, ra); |
| /* It's worth special-casing immediate loads. */ |
| if (rb == 31) { |
| tcg_gen_movi_i64(va, disp16); |
| } else { |
| tcg_gen_addi_i64(va, load_gpr(ctx, rb), disp16); |
| } |
| break; |
| |
| case 0x0A: |
| /* LDBU */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_BWX); |
| gen_load_mem(ctx, &tcg_gen_qemu_ld8u, ra, rb, disp16, 0, 0); |
| break; |
| case 0x0B: |
| /* LDQ_U */ |
| gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 0, 1); |
| break; |
| case 0x0C: |
| /* LDWU */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_BWX); |
| gen_load_mem(ctx, &tcg_gen_qemu_ld16u, ra, rb, disp16, 0, 0); |
| break; |
| case 0x0D: |
| /* STW */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_BWX); |
| gen_store_mem(ctx, &tcg_gen_qemu_st16, ra, rb, disp16, 0, 0); |
| break; |
| case 0x0E: |
| /* STB */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_BWX); |
| gen_store_mem(ctx, &tcg_gen_qemu_st8, ra, rb, disp16, 0, 0); |
| break; |
| case 0x0F: |
| /* STQ_U */ |
| gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 0, 1); |
| break; |
| |
| case 0x10: |
| vc = dest_gpr(ctx, rc); |
| vb = load_gpr_lit(ctx, rb, lit, islit); |
| |
| if (ra == 31) { |
| if (fn7 == 0x00) { |
| /* Special case ADDL as SEXTL. */ |
| tcg_gen_ext32s_i64(vc, vb); |
| break; |
| } |
| if (fn7 == 0x29) { |
| /* Special case SUBQ as NEGQ. */ |
| tcg_gen_neg_i64(vc, vb); |
| break; |
| } |
| } |
| |
| va = load_gpr(ctx, ra); |
| switch (fn7) { |
| case 0x00: |
| /* ADDL */ |
| tcg_gen_add_i64(vc, va, vb); |
| tcg_gen_ext32s_i64(vc, vc); |
| break; |
| case 0x02: |
| /* S4ADDL */ |
| tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, va, 2); |
| tcg_gen_add_i64(tmp, tmp, vb); |
| tcg_gen_ext32s_i64(vc, tmp); |
| tcg_temp_free(tmp); |
| break; |
| case 0x09: |
| /* SUBL */ |
| tcg_gen_sub_i64(vc, va, vb); |
| tcg_gen_ext32s_i64(vc, vc); |
| break; |
| case 0x0B: |
| /* S4SUBL */ |
| tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, va, 2); |
| tcg_gen_sub_i64(tmp, tmp, vb); |
| tcg_gen_ext32s_i64(vc, tmp); |
| tcg_temp_free(tmp); |
| break; |
| case 0x0F: |
| /* CMPBGE */ |
| gen_helper_cmpbge(vc, va, vb); |
| break; |
| case 0x12: |
| /* S8ADDL */ |
| tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, va, 3); |
| tcg_gen_add_i64(tmp, tmp, vb); |
| tcg_gen_ext32s_i64(vc, tmp); |
| tcg_temp_free(tmp); |
| break; |
| case 0x1B: |
| /* S8SUBL */ |
| tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, va, 3); |
| tcg_gen_sub_i64(tmp, tmp, vb); |
| tcg_gen_ext32s_i64(vc, tmp); |
| tcg_temp_free(tmp); |
| break; |
| case 0x1D: |
| /* CMPULT */ |
| tcg_gen_setcond_i64(TCG_COND_LTU, vc, va, vb); |
| break; |
| case 0x20: |
| /* ADDQ */ |
| tcg_gen_add_i64(vc, va, vb); |
| break; |
| case 0x22: |
| /* S4ADDQ */ |
| tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, va, 2); |
| tcg_gen_add_i64(vc, tmp, vb); |
| tcg_temp_free(tmp); |
| break; |
| case 0x29: |
| /* SUBQ */ |
| tcg_gen_sub_i64(vc, va, vb); |
| break; |
| case 0x2B: |
| /* S4SUBQ */ |
| tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, va, 2); |
| tcg_gen_sub_i64(vc, tmp, vb); |
| tcg_temp_free(tmp); |
| break; |
| case 0x2D: |
| /* CMPEQ */ |
| tcg_gen_setcond_i64(TCG_COND_EQ, vc, va, vb); |
| break; |
| case 0x32: |
| /* S8ADDQ */ |
| tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, va, 3); |
| tcg_gen_add_i64(vc, tmp, vb); |
| tcg_temp_free(tmp); |
| break; |
| case 0x3B: |
| /* S8SUBQ */ |
| tmp = tcg_temp_new(); |
| tcg_gen_shli_i64(tmp, va, 3); |
| tcg_gen_sub_i64(vc, tmp, vb); |
| tcg_temp_free(tmp); |
| break; |
| case 0x3D: |
| /* CMPULE */ |
| tcg_gen_setcond_i64(TCG_COND_LEU, vc, va, vb); |
| break; |
| case 0x40: |
| /* ADDL/V */ |
| tmp = tcg_temp_new(); |
| tcg_gen_ext32s_i64(tmp, va); |
| tcg_gen_ext32s_i64(vc, vb); |
| tcg_gen_add_i64(tmp, tmp, vc); |
| tcg_gen_ext32s_i64(vc, tmp); |
| gen_helper_check_overflow(cpu_env, vc, tmp); |
| tcg_temp_free(tmp); |
| break; |
| case 0x49: |
| /* SUBL/V */ |
| tmp = tcg_temp_new(); |
| tcg_gen_ext32s_i64(tmp, va); |
| tcg_gen_ext32s_i64(vc, vb); |
| tcg_gen_sub_i64(tmp, tmp, vc); |
| tcg_gen_ext32s_i64(vc, tmp); |
| gen_helper_check_overflow(cpu_env, vc, tmp); |
| tcg_temp_free(tmp); |
| break; |
| case 0x4D: |
| /* CMPLT */ |
| tcg_gen_setcond_i64(TCG_COND_LT, vc, va, vb); |
| break; |
| case 0x60: |
| /* ADDQ/V */ |
| tmp = tcg_temp_new(); |
| tmp2 = tcg_temp_new(); |
| tcg_gen_eqv_i64(tmp, va, vb); |
| tcg_gen_mov_i64(tmp2, va); |
| tcg_gen_add_i64(vc, va, vb); |
| tcg_gen_xor_i64(tmp2, tmp2, vc); |
| tcg_gen_and_i64(tmp, tmp, tmp2); |
| tcg_gen_shri_i64(tmp, tmp, 63); |
| tcg_gen_movi_i64(tmp2, 0); |
| gen_helper_check_overflow(cpu_env, tmp, tmp2); |
| tcg_temp_free(tmp); |
| tcg_temp_free(tmp2); |
| break; |
| case 0x69: |
| /* SUBQ/V */ |
| tmp = tcg_temp_new(); |
| tmp2 = tcg_temp_new(); |
| tcg_gen_xor_i64(tmp, va, vb); |
| tcg_gen_mov_i64(tmp2, va); |
| tcg_gen_sub_i64(vc, va, vb); |
| tcg_gen_xor_i64(tmp2, tmp2, vc); |
| tcg_gen_and_i64(tmp, tmp, tmp2); |
| tcg_gen_shri_i64(tmp, tmp, 63); |
| tcg_gen_movi_i64(tmp2, 0); |
| gen_helper_check_overflow(cpu_env, tmp, tmp2); |
| tcg_temp_free(tmp); |
| tcg_temp_free(tmp2); |
| break; |
| case 0x6D: |
| /* CMPLE */ |
| tcg_gen_setcond_i64(TCG_COND_LE, vc, va, vb); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x11: |
| if (fn7 == 0x20) { |
| if (rc == 31) { |
| /* Special case BIS as NOP. */ |
| break; |
| } |
| if (ra == 31) { |
| /* Special case BIS as MOV. */ |
| vc = dest_gpr(ctx, rc); |
| if (islit) { |
| tcg_gen_movi_i64(vc, lit); |
| } else { |
| tcg_gen_mov_i64(vc, load_gpr(ctx, rb)); |
| } |
| break; |
| } |
| } |
| |
| vc = dest_gpr(ctx, rc); |
| vb = load_gpr_lit(ctx, rb, lit, islit); |
| |
| if (fn7 == 0x28 && ra == 31) { |
| /* Special case ORNOT as NOT. */ |
| tcg_gen_not_i64(vc, vb); |
| break; |
| } |
| |
| va = load_gpr(ctx, ra); |
| switch (fn7) { |
| case 0x00: |
| /* AND */ |
| tcg_gen_and_i64(vc, va, vb); |
| break; |
| case 0x08: |
| /* BIC */ |
| tcg_gen_andc_i64(vc, va, vb); |
| break; |
| case 0x14: |
| /* CMOVLBS */ |
| tmp = tcg_temp_new(); |
| tcg_gen_andi_i64(tmp, va, 1); |
| tcg_gen_movcond_i64(TCG_COND_NE, vc, tmp, load_zero(ctx), |
| vb, load_gpr(ctx, rc)); |
| tcg_temp_free(tmp); |
| break; |
| case 0x16: |
| /* CMOVLBC */ |
| tmp = tcg_temp_new(); |
| tcg_gen_andi_i64(tmp, va, 1); |
| tcg_gen_movcond_i64(TCG_COND_EQ, vc, tmp, load_zero(ctx), |
| vb, load_gpr(ctx, rc)); |
| tcg_temp_free(tmp); |
| break; |
| case 0x20: |
| /* BIS */ |
| tcg_gen_or_i64(vc, va, vb); |
| break; |
| case 0x24: |
| /* CMOVEQ */ |
| tcg_gen_movcond_i64(TCG_COND_EQ, vc, va, load_zero(ctx), |
| vb, load_gpr(ctx, rc)); |
| break; |
| case 0x26: |
| /* CMOVNE */ |
| tcg_gen_movcond_i64(TCG_COND_NE, vc, va, load_zero(ctx), |
| vb, load_gpr(ctx, rc)); |
| break; |
| case 0x28: |
| /* ORNOT */ |
| tcg_gen_orc_i64(vc, va, vb); |
| break; |
| case 0x40: |
| /* XOR */ |
| tcg_gen_xor_i64(vc, va, vb); |
| break; |
| case 0x44: |
| /* CMOVLT */ |
| tcg_gen_movcond_i64(TCG_COND_LT, vc, va, load_zero(ctx), |
| vb, load_gpr(ctx, rc)); |
| break; |
| case 0x46: |
| /* CMOVGE */ |
| tcg_gen_movcond_i64(TCG_COND_GE, vc, va, load_zero(ctx), |
| vb, load_gpr(ctx, rc)); |
| break; |
| case 0x48: |
| /* EQV */ |
| tcg_gen_eqv_i64(vc, va, vb); |
| break; |
| case 0x61: |
| /* AMASK */ |
| REQUIRE_REG_31(ra); |
| { |
| uint64_t amask = ctx->tb->flags >> TB_FLAGS_AMASK_SHIFT; |
| tcg_gen_andi_i64(vc, vb, ~amask); |
| } |
| break; |
| case 0x64: |
| /* CMOVLE */ |
| tcg_gen_movcond_i64(TCG_COND_LE, vc, va, load_zero(ctx), |
| vb, load_gpr(ctx, rc)); |
| break; |
| case 0x66: |
| /* CMOVGT */ |
| tcg_gen_movcond_i64(TCG_COND_GT, vc, va, load_zero(ctx), |
| vb, load_gpr(ctx, rc)); |
| break; |
| case 0x6C: |
| /* IMPLVER */ |
| REQUIRE_REG_31(ra); |
| tcg_gen_movi_i64(vc, ctx->implver); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x12: |
| vc = dest_gpr(ctx, rc); |
| va = load_gpr(ctx, ra); |
| switch (fn7) { |
| case 0x02: |
| /* MSKBL */ |
| gen_msk_l(ctx, vc, va, rb, islit, lit, 0x01); |
| break; |
| case 0x06: |
| /* EXTBL */ |
| gen_ext_l(ctx, vc, va, rb, islit, lit, 0x01); |
| break; |
| case 0x0B: |
| /* INSBL */ |
| gen_ins_l(ctx, vc, va, rb, islit, lit, 0x01); |
| break; |
| case 0x12: |
| /* MSKWL */ |
| gen_msk_l(ctx, vc, va, rb, islit, lit, 0x03); |
| break; |
| case 0x16: |
| /* EXTWL */ |
| gen_ext_l(ctx, vc, va, rb, islit, lit, 0x03); |
| break; |
| case 0x1B: |
| /* INSWL */ |
| gen_ins_l(ctx, vc, va, rb, islit, lit, 0x03); |
| break; |
| case 0x22: |
| /* MSKLL */ |
| gen_msk_l(ctx, vc, va, rb, islit, lit, 0x0f); |
| break; |
| case 0x26: |
| /* EXTLL */ |
| gen_ext_l(ctx, vc, va, rb, islit, lit, 0x0f); |
| break; |
| case 0x2B: |
| /* INSLL */ |
| gen_ins_l(ctx, vc, va, rb, islit, lit, 0x0f); |
| break; |
| case 0x30: |
| /* ZAP */ |
| if (islit) { |
| gen_zapnoti(vc, va, ~lit); |
| } else { |
| gen_helper_zap(vc, va, load_gpr(ctx, rb)); |
| } |
| break; |
| case 0x31: |
| /* ZAPNOT */ |
| if (islit) { |
| gen_zapnoti(vc, va, lit); |
| } else { |
| gen_helper_zapnot(vc, va, load_gpr(ctx, rb)); |
| } |
| break; |
| case 0x32: |
| /* MSKQL */ |
| gen_msk_l(ctx, vc, va, rb, islit, lit, 0xff); |
| break; |
| case 0x34: |
| /* SRL */ |
| if (islit) { |
| tcg_gen_shri_i64(vc, va, lit & 0x3f); |
| } else { |
| tmp = tcg_temp_new(); |
| vb = load_gpr(ctx, rb); |
| tcg_gen_andi_i64(tmp, vb, 0x3f); |
| tcg_gen_shr_i64(vc, va, tmp); |
| tcg_temp_free(tmp); |
| } |
| break; |
| case 0x36: |
| /* EXTQL */ |
| gen_ext_l(ctx, vc, va, rb, islit, lit, 0xff); |
| break; |
| case 0x39: |
| /* SLL */ |
| if (islit) { |
| tcg_gen_shli_i64(vc, va, lit & 0x3f); |
| } else { |
| tmp = tcg_temp_new(); |
| vb = load_gpr(ctx, rb); |
| tcg_gen_andi_i64(tmp, vb, 0x3f); |
| tcg_gen_shl_i64(vc, va, tmp); |
| tcg_temp_free(tmp); |
| } |
| break; |
| case 0x3B: |
| /* INSQL */ |
| gen_ins_l(ctx, vc, va, rb, islit, lit, 0xff); |
| break; |
| case 0x3C: |
| /* SRA */ |
| if (islit) { |
| tcg_gen_sari_i64(vc, va, lit & 0x3f); |
| } else { |
| tmp = tcg_temp_new(); |
| vb = load_gpr(ctx, rb); |
| tcg_gen_andi_i64(tmp, vb, 0x3f); |
| tcg_gen_sar_i64(vc, va, tmp); |
| tcg_temp_free(tmp); |
| } |
| break; |
| case 0x52: |
| /* MSKWH */ |
| gen_msk_h(ctx, vc, va, rb, islit, lit, 0x03); |
| break; |
| case 0x57: |
| /* INSWH */ |
| gen_ins_h(ctx, vc, va, rb, islit, lit, 0x03); |
| break; |
| case 0x5A: |
| /* EXTWH */ |
| gen_ext_h(ctx, vc, va, rb, islit, lit, 0x03); |
| break; |
| case 0x62: |
| /* MSKLH */ |
| gen_msk_h(ctx, vc, va, rb, islit, lit, 0x0f); |
| break; |
| case 0x67: |
| /* INSLH */ |
| gen_ins_h(ctx, vc, va, rb, islit, lit, 0x0f); |
| break; |
| case 0x6A: |
| /* EXTLH */ |
| gen_ext_h(ctx, vc, va, rb, islit, lit, 0x0f); |
| break; |
| case 0x72: |
| /* MSKQH */ |
| gen_msk_h(ctx, vc, va, rb, islit, lit, 0xff); |
| break; |
| case 0x77: |
| /* INSQH */ |
| gen_ins_h(ctx, vc, va, rb, islit, lit, 0xff); |
| break; |
| case 0x7A: |
| /* EXTQH */ |
| gen_ext_h(ctx, vc, va, rb, islit, lit, 0xff); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x13: |
| vc = dest_gpr(ctx, rc); |
| vb = load_gpr_lit(ctx, rb, lit, islit); |
| va = load_gpr(ctx, ra); |
| switch (fn7) { |
| case 0x00: |
| /* MULL */ |
| tcg_gen_mul_i64(vc, va, vb); |
| tcg_gen_ext32s_i64(vc, vc); |
| break; |
| case 0x20: |
| /* MULQ */ |
| tcg_gen_mul_i64(vc, va, vb); |
| break; |
| case 0x30: |
| /* UMULH */ |
| tmp = tcg_temp_new(); |
| tcg_gen_mulu2_i64(tmp, vc, va, vb); |
| tcg_temp_free(tmp); |
| break; |
| case 0x40: |
| /* MULL/V */ |
| tmp = tcg_temp_new(); |
| tcg_gen_ext32s_i64(tmp, va); |
| tcg_gen_ext32s_i64(vc, vb); |
| tcg_gen_mul_i64(tmp, tmp, vc); |
| tcg_gen_ext32s_i64(vc, tmp); |
| gen_helper_check_overflow(cpu_env, vc, tmp); |
| tcg_temp_free(tmp); |
| break; |
| case 0x60: |
| /* MULQ/V */ |
| tmp = tcg_temp_new(); |
| tmp2 = tcg_temp_new(); |
| tcg_gen_muls2_i64(vc, tmp, va, vb); |
| tcg_gen_sari_i64(tmp2, vc, 63); |
| gen_helper_check_overflow(cpu_env, tmp, tmp2); |
| tcg_temp_free(tmp); |
| tcg_temp_free(tmp2); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x14: |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_FIX); |
| vc = dest_fpr(ctx, rc); |
| switch (fpfn) { /* fn11 & 0x3F */ |
| case 0x04: |
| /* ITOFS */ |
| REQUIRE_REG_31(rb); |
| t32 = tcg_temp_new_i32(); |
| va = load_gpr(ctx, ra); |
| tcg_gen_trunc_i64_i32(t32, va); |
| gen_helper_memory_to_s(vc, t32); |
| tcg_temp_free_i32(t32); |
| break; |
| case 0x0A: |
| /* SQRTF */ |
| REQUIRE_REG_31(ra); |
| vb = load_fpr(ctx, rb); |
| gen_helper_sqrtf(vc, cpu_env, vb); |
| break; |
| case 0x0B: |
| /* SQRTS */ |
| REQUIRE_REG_31(ra); |
| gen_sqrts(ctx, rb, rc, fn11); |
| break; |
| case 0x14: |
| /* ITOFF */ |
| REQUIRE_REG_31(rb); |
| t32 = tcg_temp_new_i32(); |
| va = load_gpr(ctx, ra); |
| tcg_gen_trunc_i64_i32(t32, va); |
| gen_helper_memory_to_f(vc, t32); |
| tcg_temp_free_i32(t32); |
| break; |
| case 0x24: |
| /* ITOFT */ |
| REQUIRE_REG_31(rb); |
| va = load_gpr(ctx, ra); |
| tcg_gen_mov_i64(vc, va); |
| break; |
| case 0x2A: |
| /* SQRTG */ |
| REQUIRE_REG_31(ra); |
| vb = load_fpr(ctx, rb); |
| gen_helper_sqrtg(vc, cpu_env, vb); |
| break; |
| case 0x02B: |
| /* SQRTT */ |
| REQUIRE_REG_31(ra); |
| gen_sqrtt(ctx, rb, rc, fn11); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x15: |
| /* VAX floating point */ |
| /* XXX: rounding mode and trap are ignored (!) */ |
| vc = dest_fpr(ctx, rc); |
| vb = load_fpr(ctx, rb); |
| va = load_fpr(ctx, ra); |
| switch (fpfn) { /* fn11 & 0x3F */ |
| case 0x00: |
| /* ADDF */ |
| gen_helper_addf(vc, cpu_env, va, vb); |
| break; |
| case 0x01: |
| /* SUBF */ |
| gen_helper_subf(vc, cpu_env, va, vb); |
| break; |
| case 0x02: |
| /* MULF */ |
| gen_helper_mulf(vc, cpu_env, va, vb); |
| break; |
| case 0x03: |
| /* DIVF */ |
| gen_helper_divf(vc, cpu_env, va, vb); |
| break; |
| case 0x1E: |
| /* CVTDG -- TODO */ |
| REQUIRE_REG_31(ra); |
| goto invalid_opc; |
| case 0x20: |
| /* ADDG */ |
| gen_helper_addg(vc, cpu_env, va, vb); |
| break; |
| case 0x21: |
| /* SUBG */ |
| gen_helper_subg(vc, cpu_env, va, vb); |
| break; |
| case 0x22: |
| /* MULG */ |
| gen_helper_mulg(vc, cpu_env, va, vb); |
| break; |
| case 0x23: |
| /* DIVG */ |
| gen_helper_divg(vc, cpu_env, va, vb); |
| break; |
| case 0x25: |
| /* CMPGEQ */ |
| gen_helper_cmpgeq(vc, cpu_env, va, vb); |
| break; |
| case 0x26: |
| /* CMPGLT */ |
| gen_helper_cmpglt(vc, cpu_env, va, vb); |
| break; |
| case 0x27: |
| /* CMPGLE */ |
| gen_helper_cmpgle(vc, cpu_env, va, vb); |
| break; |
| case 0x2C: |
| /* CVTGF */ |
| REQUIRE_REG_31(ra); |
| gen_helper_cvtgf(vc, cpu_env, vb); |
| break; |
| case 0x2D: |
| /* CVTGD -- TODO */ |
| REQUIRE_REG_31(ra); |
| goto invalid_opc; |
| case 0x2F: |
| /* CVTGQ */ |
| REQUIRE_REG_31(ra); |
| gen_helper_cvtgq(vc, cpu_env, vb); |
| break; |
| case 0x3C: |
| /* CVTQF */ |
| REQUIRE_REG_31(ra); |
| gen_helper_cvtqf(vc, cpu_env, vb); |
| break; |
| case 0x3E: |
| /* CVTQG */ |
| REQUIRE_REG_31(ra); |
| gen_helper_cvtqg(vc, cpu_env, vb); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x16: |
| /* IEEE floating-point */ |
| switch (fpfn) { /* fn11 & 0x3F */ |
| case 0x00: |
| /* ADDS */ |
| gen_adds(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x01: |
| /* SUBS */ |
| gen_subs(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x02: |
| /* MULS */ |
| gen_muls(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x03: |
| /* DIVS */ |
| gen_divs(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x20: |
| /* ADDT */ |
| gen_addt(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x21: |
| /* SUBT */ |
| gen_subt(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x22: |
| /* MULT */ |
| gen_mult(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x23: |
| /* DIVT */ |
| gen_divt(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x24: |
| /* CMPTUN */ |
| gen_cmptun(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x25: |
| /* CMPTEQ */ |
| gen_cmpteq(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x26: |
| /* CMPTLT */ |
| gen_cmptlt(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x27: |
| /* CMPTLE */ |
| gen_cmptle(ctx, ra, rb, rc, fn11); |
| break; |
| case 0x2C: |
| REQUIRE_REG_31(ra); |
| if (fn11 == 0x2AC || fn11 == 0x6AC) { |
| /* CVTST */ |
| gen_cvtst(ctx, rb, rc, fn11); |
| } else { |
| /* CVTTS */ |
| gen_cvtts(ctx, rb, rc, fn11); |
| } |
| break; |
| case 0x2F: |
| /* CVTTQ */ |
| REQUIRE_REG_31(ra); |
| gen_cvttq(ctx, rb, rc, fn11); |
| break; |
| case 0x3C: |
| /* CVTQS */ |
| REQUIRE_REG_31(ra); |
| gen_cvtqs(ctx, rb, rc, fn11); |
| break; |
| case 0x3E: |
| /* CVTQT */ |
| REQUIRE_REG_31(ra); |
| gen_cvtqt(ctx, rb, rc, fn11); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x17: |
| switch (fn11) { |
| case 0x010: |
| /* CVTLQ */ |
| REQUIRE_REG_31(ra); |
| vc = dest_fpr(ctx, rc); |
| vb = load_fpr(ctx, rb); |
| gen_cvtlq(vc, vb); |
| break; |
| case 0x020: |
| /* CPYS */ |
| if (rc == 31) { |
| /* Special case CPYS as FNOP. */ |
| } else { |
| vc = dest_fpr(ctx, rc); |
| va = load_fpr(ctx, ra); |
| if (ra == rb) { |
| /* Special case CPYS as FMOV. */ |
| tcg_gen_mov_i64(vc, va); |
| } else { |
| vb = load_fpr(ctx, rb); |
| gen_cpy_mask(vc, va, vb, 0, 0x8000000000000000ULL); |
| } |
| } |
| break; |
| case 0x021: |
| /* CPYSN */ |
| vc = dest_fpr(ctx, rc); |
| vb = load_fpr(ctx, rb); |
| va = load_fpr(ctx, ra); |
| gen_cpy_mask(vc, va, vb, 1, 0x8000000000000000ULL); |
| break; |
| case 0x022: |
| /* CPYSE */ |
| vc = dest_fpr(ctx, rc); |
| vb = load_fpr(ctx, rb); |
| va = load_fpr(ctx, ra); |
| gen_cpy_mask(vc, va, vb, 0, 0xFFF0000000000000ULL); |
| break; |
| case 0x024: |
| /* MT_FPCR */ |
| va = load_fpr(ctx, ra); |
| gen_helper_store_fpcr(cpu_env, va); |
| if (ctx->tb_rm == QUAL_RM_D) { |
| /* Re-do the copy of the rounding mode to fp_status |
| the next time we use dynamic rounding. */ |
| ctx->tb_rm = -1; |
| } |
| break; |
| case 0x025: |
| /* MF_FPCR */ |
| va = dest_fpr(ctx, ra); |
| gen_helper_load_fpcr(va, cpu_env); |
| break; |
| case 0x02A: |
| /* FCMOVEQ */ |
| gen_fcmov(ctx, TCG_COND_EQ, ra, rb, rc); |
| break; |
| case 0x02B: |
| /* FCMOVNE */ |
| gen_fcmov(ctx, TCG_COND_NE, ra, rb, rc); |
| break; |
| case 0x02C: |
| /* FCMOVLT */ |
| gen_fcmov(ctx, TCG_COND_LT, ra, rb, rc); |
| break; |
| case 0x02D: |
| /* FCMOVGE */ |
| gen_fcmov(ctx, TCG_COND_GE, ra, rb, rc); |
| break; |
| case 0x02E: |
| /* FCMOVLE */ |
| gen_fcmov(ctx, TCG_COND_LE, ra, rb, rc); |
| break; |
| case 0x02F: |
| /* FCMOVGT */ |
| gen_fcmov(ctx, TCG_COND_GT, ra, rb, rc); |
| break; |
| case 0x030: /* CVTQL */ |
| case 0x130: /* CVTQL/V */ |
| case 0x530: /* CVTQL/SV */ |
| REQUIRE_REG_31(ra); |
| vc = dest_fpr(ctx, rc); |
| vb = load_fpr(ctx, rb); |
| gen_helper_cvtql(vc, cpu_env, vb); |
| gen_fp_exc_raise(rc, fn11); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x18: |
| switch ((uint16_t)disp16) { |
| case 0x0000: |
| /* TRAPB */ |
| /* No-op. */ |
| break; |
| case 0x0400: |
| /* EXCB */ |
| /* No-op. */ |
| break; |
| case 0x4000: |
| /* MB */ |
| /* No-op */ |
| break; |
| case 0x4400: |
| /* WMB */ |
| /* No-op */ |
| break; |
| case 0x8000: |
| /* FETCH */ |
| /* No-op */ |
| break; |
| case 0xA000: |
| /* FETCH_M */ |
| /* No-op */ |
| break; |
| case 0xC000: |
| /* RPCC */ |
| va = dest_gpr(ctx, ra); |
| if (ctx->tb->cflags & CF_USE_ICOUNT) { |
| gen_io_start(); |
| gen_helper_load_pcc(va, cpu_env); |
| gen_io_end(); |
| ret = EXIT_PC_STALE; |
| } else { |
| gen_helper_load_pcc(va, cpu_env); |
| } |
| break; |
| case 0xE000: |
| /* RC */ |
| gen_rx(ra, 0); |
| break; |
| case 0xE800: |
| /* ECB */ |
| break; |
| case 0xF000: |
| /* RS */ |
| gen_rx(ra, 1); |
| break; |
| case 0xF800: |
| /* WH64 */ |
| /* No-op */ |
| break; |
| case 0xFC00: |
| /* WH64EN */ |
| /* No-op */ |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x19: |
| /* HW_MFPR (PALcode) */ |
| #ifndef CONFIG_USER_ONLY |
| REQUIRE_TB_FLAG(TB_FLAGS_PAL_MODE); |
| va = dest_gpr(ctx, ra); |
| ret = gen_mfpr(ctx, va, insn & 0xffff); |
| break; |
| #else |
| goto invalid_opc; |
| #endif |
| |
| case 0x1A: |
| /* JMP, JSR, RET, JSR_COROUTINE. These only differ by the branch |
| prediction stack action, which of course we don't implement. */ |
| vb = load_gpr(ctx, rb); |
| tcg_gen_andi_i64(cpu_pc, vb, ~3); |
| if (ra != 31) { |
| tcg_gen_movi_i64(cpu_ir[ra], ctx->pc); |
| } |
| ret = EXIT_PC_UPDATED; |
| break; |
| |
| case 0x1B: |
| /* HW_LD (PALcode) */ |
| #ifndef CONFIG_USER_ONLY |
| REQUIRE_TB_FLAG(TB_FLAGS_PAL_MODE); |
| { |
| TCGv addr = tcg_temp_new(); |
| vb = load_gpr(ctx, rb); |
| va = dest_gpr(ctx, ra); |
| |
| tcg_gen_addi_i64(addr, vb, disp12); |
| switch ((insn >> 12) & 0xF) { |
| case 0x0: |
| /* Longword physical access (hw_ldl/p) */ |
| gen_helper_ldl_phys(va, cpu_env, addr); |
| break; |
| case 0x1: |
| /* Quadword physical access (hw_ldq/p) */ |
| gen_helper_ldq_phys(va, cpu_env, addr); |
| break; |
| case 0x2: |
| /* Longword physical access with lock (hw_ldl_l/p) */ |
| gen_helper_ldl_l_phys(va, cpu_env, addr); |
| break; |
| case 0x3: |
| /* Quadword physical access with lock (hw_ldq_l/p) */ |
| gen_helper_ldq_l_phys(va, cpu_env, addr); |
| break; |
| case 0x4: |
| /* Longword virtual PTE fetch (hw_ldl/v) */ |
| goto invalid_opc; |
| case 0x5: |
| /* Quadword virtual PTE fetch (hw_ldq/v) */ |
| goto invalid_opc; |
| break; |
| case 0x6: |
| /* Incpu_ir[ra]id */ |
| goto invalid_opc; |
| case 0x7: |
| /* Incpu_ir[ra]id */ |
| goto invalid_opc; |
| case 0x8: |
| /* Longword virtual access (hw_ldl) */ |
| goto invalid_opc; |
| case 0x9: |
| /* Quadword virtual access (hw_ldq) */ |
| goto invalid_opc; |
| case 0xA: |
| /* Longword virtual access with protection check (hw_ldl/w) */ |
| tcg_gen_qemu_ld_i64(va, addr, MMU_KERNEL_IDX, MO_LESL); |
| break; |
| case 0xB: |
| /* Quadword virtual access with protection check (hw_ldq/w) */ |
| tcg_gen_qemu_ld_i64(va, addr, MMU_KERNEL_IDX, MO_LEQ); |
| break; |
| case 0xC: |
| /* Longword virtual access with alt access mode (hw_ldl/a)*/ |
| goto invalid_opc; |
| case 0xD: |
| /* Quadword virtual access with alt access mode (hw_ldq/a) */ |
| goto invalid_opc; |
| case 0xE: |
| /* Longword virtual access with alternate access mode and |
| protection checks (hw_ldl/wa) */ |
| tcg_gen_qemu_ld_i64(va, addr, MMU_USER_IDX, MO_LESL); |
| break; |
| case 0xF: |
| /* Quadword virtual access with alternate access mode and |
| protection checks (hw_ldq/wa) */ |
| tcg_gen_qemu_ld_i64(va, addr, MMU_USER_IDX, MO_LEQ); |
| break; |
| } |
| tcg_temp_free(addr); |
| break; |
| } |
| #else |
| goto invalid_opc; |
| #endif |
| |
| case 0x1C: |
| vc = dest_gpr(ctx, rc); |
| if (fn7 == 0x70) { |
| /* FTOIT */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_FIX); |
| REQUIRE_REG_31(rb); |
| va = load_fpr(ctx, ra); |
| tcg_gen_mov_i64(vc, va); |
| break; |
| } else if (fn7 == 0x78) { |
| /* FTOIS */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_FIX); |
| REQUIRE_REG_31(rb); |
| t32 = tcg_temp_new_i32(); |
| va = load_fpr(ctx, ra); |
| gen_helper_s_to_memory(t32, va); |
| tcg_gen_ext_i32_i64(vc, t32); |
| tcg_temp_free_i32(t32); |
| break; |
| } |
| |
| vb = load_gpr_lit(ctx, rb, lit, islit); |
| switch (fn7) { |
| case 0x00: |
| /* SEXTB */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_BWX); |
| REQUIRE_REG_31(ra); |
| tcg_gen_ext8s_i64(vc, vb); |
| break; |
| case 0x01: |
| /* SEXTW */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_BWX); |
| REQUIRE_REG_31(ra); |
| tcg_gen_ext16s_i64(vc, vb); |
| break; |
| case 0x30: |
| /* CTPOP */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_CIX); |
| REQUIRE_REG_31(ra); |
| REQUIRE_NO_LIT; |
| gen_helper_ctpop(vc, vb); |
| break; |
| case 0x31: |
| /* PERR */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| REQUIRE_NO_LIT; |
| va = load_gpr(ctx, ra); |
| gen_helper_perr(vc, va, vb); |
| break; |
| case 0x32: |
| /* CTLZ */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_CIX); |
| REQUIRE_REG_31(ra); |
| REQUIRE_NO_LIT; |
| gen_helper_ctlz(vc, vb); |
| break; |
| case 0x33: |
| /* CTTZ */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_CIX); |
| REQUIRE_REG_31(ra); |
| REQUIRE_NO_LIT; |
| gen_helper_cttz(vc, vb); |
| break; |
| case 0x34: |
| /* UNPKBW */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| REQUIRE_REG_31(ra); |
| REQUIRE_NO_LIT; |
| gen_helper_unpkbw(vc, vb); |
| break; |
| case 0x35: |
| /* UNPKBL */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| REQUIRE_REG_31(ra); |
| REQUIRE_NO_LIT; |
| gen_helper_unpkbl(vc, vb); |
| break; |
| case 0x36: |
| /* PKWB */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| REQUIRE_REG_31(ra); |
| REQUIRE_NO_LIT; |
| gen_helper_pkwb(vc, vb); |
| break; |
| case 0x37: |
| /* PKLB */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| REQUIRE_REG_31(ra); |
| REQUIRE_NO_LIT; |
| gen_helper_pklb(vc, vb); |
| break; |
| case 0x38: |
| /* MINSB8 */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| va = load_gpr(ctx, ra); |
| gen_helper_minsb8(vc, va, vb); |
| break; |
| case 0x39: |
| /* MINSW4 */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| va = load_gpr(ctx, ra); |
| gen_helper_minsw4(vc, va, vb); |
| break; |
| case 0x3A: |
| /* MINUB8 */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| va = load_gpr(ctx, ra); |
| gen_helper_minub8(vc, va, vb); |
| break; |
| case 0x3B: |
| /* MINUW4 */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| va = load_gpr(ctx, ra); |
| gen_helper_minuw4(vc, va, vb); |
| break; |
| case 0x3C: |
| /* MAXUB8 */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| va = load_gpr(ctx, ra); |
| gen_helper_maxub8(vc, va, vb); |
| break; |
| case 0x3D: |
| /* MAXUW4 */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| va = load_gpr(ctx, ra); |
| gen_helper_maxuw4(vc, va, vb); |
| break; |
| case 0x3E: |
| /* MAXSB8 */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| va = load_gpr(ctx, ra); |
| gen_helper_maxsb8(vc, va, vb); |
| break; |
| case 0x3F: |
| /* MAXSW4 */ |
| REQUIRE_TB_FLAG(TB_FLAGS_AMASK_MVI); |
| va = load_gpr(ctx, ra); |
| gen_helper_maxsw4(vc, va, vb); |
| break; |
| default: |
| goto invalid_opc; |
| } |
| break; |
| |
| case 0x1D: |
| /* HW_MTPR (PALcode) */ |
| #ifndef CONFIG_USER_ONLY |
| REQUIRE_TB_FLAG(TB_FLAGS_PAL_MODE); |
| vb = load_gpr(ctx, rb); |
| ret = gen_mtpr(ctx, vb, insn & 0xffff); |
| break; |
| #else |
| goto invalid_opc; |
| #endif |
| |
| case 0x1E: |
| /* HW_RET (PALcode) */ |
| #ifndef CONFIG_USER_ONLY |
| REQUIRE_TB_FLAG(TB_FLAGS_PAL_MODE); |
| if (rb == 31) { |
| /* Pre-EV6 CPUs interpreted this as HW_REI, loading the return |
| address from EXC_ADDR. This turns out to be useful for our |
| emulation PALcode, so continue to accept it. */ |
| tmp = tcg_temp_new(); |
| tcg_gen_ld_i64(tmp, cpu_env, offsetof(CPUAlphaState, exc_addr)); |
| gen_helper_hw_ret(cpu_env, tmp); |
| tcg_temp_free(tmp); |
| } else { |
| gen_helper_hw_ret(cpu_env, load_gpr(ctx, rb)); |
| } |
| ret = EXIT_PC_UPDATED; |
| break; |
| #else |
| goto invalid_opc; |
| #endif |
| |
| case 0x1F: |
| /* HW_ST (PALcode) */ |
| #ifndef CONFIG_USER_ONLY |
| REQUIRE_TB_FLAG(TB_FLAGS_PAL_MODE); |
| { |
| TCGv addr = tcg_temp_new(); |
| va = load_gpr(ctx, ra); |
| vb = load_gpr(ctx, rb); |
| |
| tcg_gen_addi_i64(addr, vb, disp12); |
| switch ((insn >> 12) & 0xF) { |
| case 0x0: |
| /* Longword physical access */ |
| gen_helper_stl_phys(cpu_env, addr, va); |
| break; |
| case 0x1: |
| /* Quadword physical access */ |
| gen_helper_stq_phys(cpu_env, addr, va); |
| break; |
| case 0x2: |
| /* Longword physical access with lock */ |
| gen_helper_stl_c_phys(dest_gpr(ctx, ra), cpu_env, addr, va); |
| break; |
| case 0x3: |
| /* Quadword physical access with lock */ |
| gen_helper_stq_c_phys(dest_gpr(ctx, ra), cpu_env, addr, va); |
| break; |
| case 0x4: |
| /* Longword virtual access */ |
| goto invalid_opc; |
| case 0x5: |
| /* Quadword virtual access */ |
| goto invalid_opc; |
| case 0x6: |
| /* Invalid */ |
| goto invalid_opc; |
| case 0x7: |
| /* Invalid */ |
| goto invalid_opc; |
| case 0x8: |
| /* Invalid */ |
| goto invalid_opc; |
| case 0x9: |
| /* Invalid */ |
| goto invalid_opc; |
| case 0xA: |
| /* Invalid */ |
| goto invalid_opc; |
| case 0xB: |
| /* Invalid */ |
| goto invalid_opc; |
| case 0xC: |
| /* Longword virtual access with alternate access mode */ |
| goto invalid_opc; |
| case 0xD: |
| /* Quadword virtual access with alternate access mode */ |
| goto invalid_opc; |
| case 0xE: |
| /* Invalid */ |
| goto invalid_opc; |
| case 0xF: |
| /* Invalid */ |
| goto invalid_opc; |
| } |
| tcg_temp_free(addr); |
| break; |
| } |
| #else |
| goto invalid_opc; |
| #endif |
| case 0x20: |
| /* LDF */ |
| gen_load_mem(ctx, &gen_qemu_ldf, ra, rb, disp16, 1, 0); |
| break; |
| case 0x21: |
| /* LDG */ |
| gen_load_mem(ctx, &gen_qemu_ldg, ra, rb, disp16, 1, 0); |
| break; |
| case 0x22: |
| /* LDS */ |
| gen_load_mem(ctx, &gen_qemu_lds, ra, rb, disp16, 1, 0); |
| break; |
| case 0x23: |
| /* LDT */ |
| gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 1, 0); |
| break; |
| case 0x24: |
| /* STF */ |
| gen_store_mem(ctx, &gen_qemu_stf, ra, rb, disp16, 1, 0); |
| break; |
| case 0x25: |
| /* STG */ |
| gen_store_mem(ctx, &gen_qemu_stg, ra, rb, disp16, 1, 0); |
| break; |
| case 0x26: |
| /* STS */ |
| gen_store_mem(ctx, &gen_qemu_sts, ra, rb, disp16, 1, 0); |
| break; |
| case 0x27: |
| /* STT */ |
| gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 1, 0); |
| break; |
| case 0x28: |
| /* LDL */ |
| gen_load_mem(ctx, &tcg_gen_qemu_ld32s, ra, rb, disp16, 0, 0); |
| break; |
| case 0x29: |
| /* LDQ */ |
| gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 0, 0); |
| break; |
| case 0x2A: |
| /* LDL_L */ |
| gen_load_mem(ctx, &gen_qemu_ldl_l, ra, rb, disp16, 0, 0); |
| break; |
| case 0x2B: |
| /* LDQ_L */ |
| gen_load_mem(ctx, &gen_qemu_ldq_l, ra, rb, disp16, 0, 0); |
| break; |
| case 0x2C: |
| /* STL */ |
| gen_store_mem(ctx, &tcg_gen_qemu_st32, ra, rb, disp16, 0, 0); |
| break; |
| case 0x2D: |
| /* STQ */ |
| gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 0, 0); |
| break; |
| case 0x2E: |
| /* STL_C */ |
| ret = gen_store_conditional(ctx, ra, rb, disp16, 0); |
| break; |
| case 0x2F: |
| /* STQ_C */ |
| ret = gen_store_conditional(ctx, ra, rb, disp16, 1); |
| break; |
| case 0x30: |
| /* BR */ |
| ret = gen_bdirect(ctx, ra, disp21); |
| break; |
| case 0x31: /* FBEQ */ |
| ret = gen_fbcond(ctx, TCG_COND_EQ, ra, disp21); |
| break; |
| case 0x32: /* FBLT */ |
| ret = gen_fbcond(ctx, TCG_COND_LT, ra, disp21); |
| break; |
| case 0x33: /* FBLE */ |
| ret = gen_fbcond(ctx, TCG_COND_LE, ra, disp21); |
| break; |
| case 0x34: |
| /* BSR */ |
| ret = gen_bdirect(ctx, ra, disp21); |
| break; |
| case 0x35: /* FBNE */ |
| ret = gen_fbcond(ctx, TCG_COND_NE, ra, disp21); |
| break; |
| case 0x36: /* FBGE */ |
| ret = gen_fbcond(ctx, TCG_COND_GE, ra, disp21); |
| break; |
| case 0x37: /* FBGT */ |
| ret = gen_fbcond(ctx, TCG_COND_GT, ra, disp21); |
| break; |
| case 0x38: |
| /* BLBC */ |
| ret = gen_bcond(ctx, TCG_COND_EQ, ra, disp21, 1); |
| break; |
| case 0x39: |
| /* BEQ */ |
| ret = gen_bcond(ctx, TCG_COND_EQ, ra, disp21, 0); |
| break; |
| case 0x3A: |
| /* BLT */ |
| ret = gen_bcond(ctx, TCG_COND_LT, ra, disp21, 0); |
| break; |
| case 0x3B: |
| /* BLE */ |
| ret = gen_bcond(ctx, TCG_COND_LE, ra, disp21, 0); |
| break; |
| case 0x3C: |
| /* BLBS */ |
| ret = gen_bcond(ctx, TCG_COND_NE, ra, disp21, 1); |
| break; |
| case 0x3D: |
| /* BNE */ |
| ret = gen_bcond(ctx, TCG_COND_NE, ra, disp21, 0); |
| break; |
| case 0x3E: |
| /* BGE */ |
| ret = gen_bcond(ctx, TCG_COND_GE, ra, disp21, 0); |
| break; |
| case 0x3F: |
| /* BGT */ |
| ret = gen_bcond(ctx, TCG_COND_GT, ra, disp21, 0); |
| break; |
| invalid_opc: |
| ret = gen_invalid(ctx); |
| break; |
| } |
| |
| return ret; |
| } |
| |
| static inline void gen_intermediate_code_internal(AlphaCPU *cpu, |
| TranslationBlock *tb, |
| bool search_pc) |
| { |
| CPUState *cs = CPU(cpu); |
| CPUAlphaState *env = &cpu->env; |
| DisasContext ctx, *ctxp = &ctx; |
| target_ulong pc_start; |
| target_ulong pc_mask; |
| uint32_t insn; |
| CPUBreakpoint *bp; |
| int j, lj = -1; |
| ExitStatus ret; |
| int num_insns; |
| int max_insns; |
| |
| pc_start = tb->pc; |
| |
| ctx.tb = tb; |
| ctx.pc = pc_start; |
| ctx.mem_idx = cpu_mmu_index(env); |
| ctx.implver = env->implver; |
| ctx.singlestep_enabled = cs->singlestep_enabled; |
| |
| /* ??? Every TB begins with unset rounding mode, to be initialized on |
| the first fp insn of the TB. Alternately we could define a proper |
| default for every TB (e.g. QUAL_RM_N or QUAL_RM_D) and make sure |
| to reset the FP_STATUS to that default at the end of any TB that |
| changes the default. We could even (gasp) dynamiclly figure out |
| what default would be most efficient given the running program. */ |
| ctx.tb_rm = -1; |
| /* Similarly for flush-to-zero. */ |
| ctx.tb_ftz = -1; |
| |
| num_insns = 0; |
| max_insns = tb->cflags & CF_COUNT_MASK; |
| if (max_insns == 0) { |
| max_insns = CF_COUNT_MASK; |
| } |
| |
| if (in_superpage(&ctx, pc_start)) { |
| pc_mask = (1ULL << 41) - 1; |
| } else { |
| pc_mask = ~TARGET_PAGE_MASK; |
| } |
| |
| gen_tb_start(tb); |
| do { |
| if (unlikely(!QTAILQ_EMPTY(&cs->breakpoints))) { |
| QTAILQ_FOREACH(bp, &cs->breakpoints, entry) { |
| if (bp->pc == ctx.pc) { |
| gen_excp(&ctx, EXCP_DEBUG, 0); |
| break; |
| } |
| } |
| } |
| if (search_pc) { |
| j = tcg_op_buf_count(); |
| if (lj < j) { |
| lj++; |
| while (lj < j) { |
| tcg_ctx.gen_opc_instr_start[lj++] = 0; |
| } |
| } |
| tcg_ctx.gen_opc_pc[lj] = ctx.pc; |
| tcg_ctx.gen_opc_instr_start[lj] = 1; |
| tcg_ctx.gen_opc_icount[lj] = num_insns; |
| } |
| if (num_insns + 1 == max_insns && (tb->cflags & CF_LAST_IO)) { |
| gen_io_start(); |
| } |
| insn = cpu_ldl_code(env, ctx.pc); |
| num_insns++; |
| |
| if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP | CPU_LOG_TB_OP_OPT))) { |
| tcg_gen_debug_insn_start(ctx.pc); |
| } |
| |
| TCGV_UNUSED_I64(ctx.zero); |
| TCGV_UNUSED_I64(ctx.sink); |
| TCGV_UNUSED_I64(ctx.lit); |
| |
| ctx.pc += 4; |
| ret = translate_one(ctxp, insn); |
| |
| if (!TCGV_IS_UNUSED_I64(ctx.sink)) { |
| tcg_gen_discard_i64(ctx.sink); |
| tcg_temp_free(ctx.sink); |
| } |
| if (!TCGV_IS_UNUSED_I64(ctx.zero)) { |
| tcg_temp_free(ctx.zero); |
| } |
| if (!TCGV_IS_UNUSED_I64(ctx.lit)) { |
| tcg_temp_free(ctx.lit); |
| } |
| |
| /* If we reach a page boundary, are single stepping, |
| or exhaust instruction count, stop generation. */ |
| if (ret == NO_EXIT |
| && ((ctx.pc & pc_mask) == 0 |
| || tcg_op_buf_full() |
| || num_insns >= max_insns |
| || singlestep |
| || ctx.singlestep_enabled)) { |
| ret = EXIT_PC_STALE; |
| } |
| } while (ret == NO_EXIT); |
| |
| if (tb->cflags & CF_LAST_IO) { |
| gen_io_end(); |
| } |
| |
| switch (ret) { |
| case EXIT_GOTO_TB: |
| case EXIT_NORETURN: |
| break; |
| case EXIT_PC_STALE: |
| tcg_gen_movi_i64(cpu_pc, ctx.pc); |
| /* FALLTHRU */ |
| case EXIT_PC_UPDATED: |
| if (ctx.singlestep_enabled) { |
| gen_excp_1(EXCP_DEBUG, 0); |
| } else { |
| tcg_gen_exit_tb(0); |
| } |
| break; |
| default: |
| abort(); |
| } |
| |
| gen_tb_end(tb, num_insns); |
| |
| if (search_pc) { |
| j = tcg_op_buf_count(); |
| lj++; |
| while (lj <= j) { |
| tcg_ctx.gen_opc_instr_start[lj++] = 0; |
| } |
| } else { |
| tb->size = ctx.pc - pc_start; |
| tb->icount = num_insns; |
| } |
| |
| #ifdef DEBUG_DISAS |
| if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) { |
| qemu_log("IN: %s\n", lookup_symbol(pc_start)); |
| log_target_disas(cs, pc_start, ctx.pc - pc_start, 1); |
| qemu_log("\n"); |
| } |
| #endif |
| } |
| |
| void gen_intermediate_code (CPUAlphaState *env, struct TranslationBlock *tb) |
| { |
| gen_intermediate_code_internal(alpha_env_get_cpu(env), tb, false); |
| } |
| |
| void gen_intermediate_code_pc (CPUAlphaState *env, struct TranslationBlock *tb) |
| { |
| gen_intermediate_code_internal(alpha_env_get_cpu(env), tb, true); |
| } |
| |
| void restore_state_to_opc(CPUAlphaState *env, TranslationBlock *tb, int pc_pos) |
| { |
| env->pc = tcg_ctx.gen_opc_pc[pc_pos]; |
| } |