| /* |
| * RX translation |
| * |
| * Copyright (c) 2019 Yoshinori Sato |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms and conditions of the GNU General Public License, |
| * version 2 or later, as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| * |
| * You should have received a copy of the GNU General Public License along with |
| * this program. If not, see <http://www.gnu.org/licenses/>. |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "qemu/bswap.h" |
| #include "qemu/qemu-print.h" |
| #include "cpu.h" |
| #include "exec/exec-all.h" |
| #include "tcg/tcg-op.h" |
| #include "exec/cpu_ldst.h" |
| #include "exec/helper-proto.h" |
| #include "exec/helper-gen.h" |
| #include "exec/translator.h" |
| #include "exec/log.h" |
| |
| typedef struct DisasContext { |
| DisasContextBase base; |
| CPURXState *env; |
| uint32_t pc; |
| } DisasContext; |
| |
| typedef struct DisasCompare { |
| TCGv value; |
| TCGv temp; |
| TCGCond cond; |
| } DisasCompare; |
| |
| const char *rx_crname(uint8_t cr) |
| { |
| static const char *cr_names[] = { |
| "psw", "pc", "usp", "fpsw", "", "", "", "", |
| "bpsw", "bpc", "isp", "fintv", "intb", "", "", "" |
| }; |
| if (cr >= ARRAY_SIZE(cr_names)) { |
| return "illegal"; |
| } |
| return cr_names[cr]; |
| } |
| |
| /* Target-specific values for dc->base.is_jmp. */ |
| #define DISAS_JUMP DISAS_TARGET_0 |
| #define DISAS_UPDATE DISAS_TARGET_1 |
| #define DISAS_EXIT DISAS_TARGET_2 |
| |
| /* global register indexes */ |
| static TCGv cpu_regs[16]; |
| static TCGv cpu_psw_o, cpu_psw_s, cpu_psw_z, cpu_psw_c; |
| static TCGv cpu_psw_i, cpu_psw_pm, cpu_psw_u, cpu_psw_ipl; |
| static TCGv cpu_usp, cpu_fpsw, cpu_bpsw, cpu_bpc, cpu_isp; |
| static TCGv cpu_fintv, cpu_intb, cpu_pc; |
| static TCGv_i64 cpu_acc; |
| |
| #define cpu_sp cpu_regs[0] |
| |
| #include "exec/gen-icount.h" |
| |
| /* decoder helper */ |
| static uint32_t decode_load_bytes(DisasContext *ctx, uint32_t insn, |
| int i, int n) |
| { |
| while (++i <= n) { |
| uint8_t b = cpu_ldub_code(ctx->env, ctx->base.pc_next++); |
| insn |= b << (32 - i * 8); |
| } |
| return insn; |
| } |
| |
| static uint32_t li(DisasContext *ctx, int sz) |
| { |
| int32_t tmp, addr; |
| CPURXState *env = ctx->env; |
| addr = ctx->base.pc_next; |
| |
| tcg_debug_assert(sz < 4); |
| switch (sz) { |
| case 1: |
| ctx->base.pc_next += 1; |
| return cpu_ldsb_code(env, addr); |
| case 2: |
| ctx->base.pc_next += 2; |
| return cpu_ldsw_code(env, addr); |
| case 3: |
| ctx->base.pc_next += 3; |
| tmp = cpu_ldsb_code(env, addr + 2) << 16; |
| tmp |= cpu_lduw_code(env, addr) & 0xffff; |
| return tmp; |
| case 0: |
| ctx->base.pc_next += 4; |
| return cpu_ldl_code(env, addr); |
| } |
| return 0; |
| } |
| |
| static int bdsp_s(DisasContext *ctx, int d) |
| { |
| /* |
| * 0 -> 8 |
| * 1 -> 9 |
| * 2 -> 10 |
| * 3 -> 3 |
| * : |
| * 7 -> 7 |
| */ |
| if (d < 3) { |
| d += 8; |
| } |
| return d; |
| } |
| |
| /* Include the auto-generated decoder. */ |
| #include "decode-insns.c.inc" |
| |
| void rx_cpu_dump_state(CPUState *cs, FILE *f, int flags) |
| { |
| RXCPU *cpu = RX_CPU(cs); |
| CPURXState *env = &cpu->env; |
| int i; |
| uint32_t psw; |
| |
| psw = rx_cpu_pack_psw(env); |
| qemu_fprintf(f, "pc=0x%08x psw=0x%08x\n", |
| env->pc, psw); |
| for (i = 0; i < 16; i += 4) { |
| qemu_fprintf(f, "r%d=0x%08x r%d=0x%08x r%d=0x%08x r%d=0x%08x\n", |
| i, env->regs[i], i + 1, env->regs[i + 1], |
| i + 2, env->regs[i + 2], i + 3, env->regs[i + 3]); |
| } |
| } |
| |
| static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest) |
| { |
| if (translator_use_goto_tb(&dc->base, dest)) { |
| tcg_gen_goto_tb(n); |
| tcg_gen_movi_i32(cpu_pc, dest); |
| tcg_gen_exit_tb(dc->base.tb, n); |
| } else { |
| tcg_gen_movi_i32(cpu_pc, dest); |
| if (dc->base.singlestep_enabled) { |
| gen_helper_debug(cpu_env); |
| } else { |
| tcg_gen_lookup_and_goto_ptr(); |
| } |
| } |
| dc->base.is_jmp = DISAS_NORETURN; |
| } |
| |
| /* generic load wrapper */ |
| static inline void rx_gen_ld(unsigned int size, TCGv reg, TCGv mem) |
| { |
| tcg_gen_qemu_ld_i32(reg, mem, 0, size | MO_SIGN | MO_TE); |
| } |
| |
| /* unsigned load wrapper */ |
| static inline void rx_gen_ldu(unsigned int size, TCGv reg, TCGv mem) |
| { |
| tcg_gen_qemu_ld_i32(reg, mem, 0, size | MO_TE); |
| } |
| |
| /* generic store wrapper */ |
| static inline void rx_gen_st(unsigned int size, TCGv reg, TCGv mem) |
| { |
| tcg_gen_qemu_st_i32(reg, mem, 0, size | MO_TE); |
| } |
| |
| /* [ri, rb] */ |
| static inline void rx_gen_regindex(DisasContext *ctx, TCGv mem, |
| int size, int ri, int rb) |
| { |
| tcg_gen_shli_i32(mem, cpu_regs[ri], size); |
| tcg_gen_add_i32(mem, mem, cpu_regs[rb]); |
| } |
| |
| /* dsp[reg] */ |
| static inline TCGv rx_index_addr(DisasContext *ctx, TCGv mem, |
| int ld, int size, int reg) |
| { |
| uint32_t dsp; |
| |
| tcg_debug_assert(ld < 3); |
| switch (ld) { |
| case 0: |
| return cpu_regs[reg]; |
| case 1: |
| dsp = cpu_ldub_code(ctx->env, ctx->base.pc_next) << size; |
| tcg_gen_addi_i32(mem, cpu_regs[reg], dsp); |
| ctx->base.pc_next += 1; |
| return mem; |
| case 2: |
| dsp = cpu_lduw_code(ctx->env, ctx->base.pc_next) << size; |
| tcg_gen_addi_i32(mem, cpu_regs[reg], dsp); |
| ctx->base.pc_next += 2; |
| return mem; |
| } |
| return NULL; |
| } |
| |
| static inline MemOp mi_to_mop(unsigned mi) |
| { |
| static const MemOp mop[5] = { MO_SB, MO_SW, MO_UL, MO_UW, MO_UB }; |
| tcg_debug_assert(mi < 5); |
| return mop[mi]; |
| } |
| |
| /* load source operand */ |
| static inline TCGv rx_load_source(DisasContext *ctx, TCGv mem, |
| int ld, int mi, int rs) |
| { |
| TCGv addr; |
| MemOp mop; |
| if (ld < 3) { |
| mop = mi_to_mop(mi); |
| addr = rx_index_addr(ctx, mem, ld, mop & MO_SIZE, rs); |
| tcg_gen_qemu_ld_i32(mem, addr, 0, mop | MO_TE); |
| return mem; |
| } else { |
| return cpu_regs[rs]; |
| } |
| } |
| |
| /* Processor mode check */ |
| static int is_privileged(DisasContext *ctx, int is_exception) |
| { |
| if (FIELD_EX32(ctx->base.tb->flags, PSW, PM)) { |
| if (is_exception) { |
| gen_helper_raise_privilege_violation(cpu_env); |
| } |
| return 0; |
| } else { |
| return 1; |
| } |
| } |
| |
| /* generate QEMU condition */ |
| static void psw_cond(DisasCompare *dc, uint32_t cond) |
| { |
| tcg_debug_assert(cond < 16); |
| switch (cond) { |
| case 0: /* z */ |
| dc->cond = TCG_COND_EQ; |
| dc->value = cpu_psw_z; |
| break; |
| case 1: /* nz */ |
| dc->cond = TCG_COND_NE; |
| dc->value = cpu_psw_z; |
| break; |
| case 2: /* c */ |
| dc->cond = TCG_COND_NE; |
| dc->value = cpu_psw_c; |
| break; |
| case 3: /* nc */ |
| dc->cond = TCG_COND_EQ; |
| dc->value = cpu_psw_c; |
| break; |
| case 4: /* gtu (C& ~Z) == 1 */ |
| case 5: /* leu (C& ~Z) == 0 */ |
| tcg_gen_setcondi_i32(TCG_COND_NE, dc->temp, cpu_psw_z, 0); |
| tcg_gen_and_i32(dc->temp, dc->temp, cpu_psw_c); |
| dc->cond = (cond == 4) ? TCG_COND_NE : TCG_COND_EQ; |
| dc->value = dc->temp; |
| break; |
| case 6: /* pz (S == 0) */ |
| dc->cond = TCG_COND_GE; |
| dc->value = cpu_psw_s; |
| break; |
| case 7: /* n (S == 1) */ |
| dc->cond = TCG_COND_LT; |
| dc->value = cpu_psw_s; |
| break; |
| case 8: /* ge (S^O)==0 */ |
| case 9: /* lt (S^O)==1 */ |
| tcg_gen_xor_i32(dc->temp, cpu_psw_o, cpu_psw_s); |
| dc->cond = (cond == 8) ? TCG_COND_GE : TCG_COND_LT; |
| dc->value = dc->temp; |
| break; |
| case 10: /* gt ((S^O)|Z)==0 */ |
| case 11: /* le ((S^O)|Z)==1 */ |
| tcg_gen_xor_i32(dc->temp, cpu_psw_o, cpu_psw_s); |
| tcg_gen_sari_i32(dc->temp, dc->temp, 31); |
| tcg_gen_andc_i32(dc->temp, cpu_psw_z, dc->temp); |
| dc->cond = (cond == 10) ? TCG_COND_NE : TCG_COND_EQ; |
| dc->value = dc->temp; |
| break; |
| case 12: /* o */ |
| dc->cond = TCG_COND_LT; |
| dc->value = cpu_psw_o; |
| break; |
| case 13: /* no */ |
| dc->cond = TCG_COND_GE; |
| dc->value = cpu_psw_o; |
| break; |
| case 14: /* always true */ |
| dc->cond = TCG_COND_ALWAYS; |
| dc->value = dc->temp; |
| break; |
| case 15: /* always false */ |
| dc->cond = TCG_COND_NEVER; |
| dc->value = dc->temp; |
| break; |
| } |
| } |
| |
| static void move_from_cr(TCGv ret, int cr, uint32_t pc) |
| { |
| TCGv z = tcg_const_i32(0); |
| switch (cr) { |
| case 0: /* PSW */ |
| gen_helper_pack_psw(ret, cpu_env); |
| break; |
| case 1: /* PC */ |
| tcg_gen_movi_i32(ret, pc); |
| break; |
| case 2: /* USP */ |
| tcg_gen_movcond_i32(TCG_COND_NE, ret, |
| cpu_psw_u, z, cpu_sp, cpu_usp); |
| break; |
| case 3: /* FPSW */ |
| tcg_gen_mov_i32(ret, cpu_fpsw); |
| break; |
| case 8: /* BPSW */ |
| tcg_gen_mov_i32(ret, cpu_bpsw); |
| break; |
| case 9: /* BPC */ |
| tcg_gen_mov_i32(ret, cpu_bpc); |
| break; |
| case 10: /* ISP */ |
| tcg_gen_movcond_i32(TCG_COND_EQ, ret, |
| cpu_psw_u, z, cpu_sp, cpu_isp); |
| break; |
| case 11: /* FINTV */ |
| tcg_gen_mov_i32(ret, cpu_fintv); |
| break; |
| case 12: /* INTB */ |
| tcg_gen_mov_i32(ret, cpu_intb); |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, "Unimplement control register %d", cr); |
| /* Unimplement registers return 0 */ |
| tcg_gen_movi_i32(ret, 0); |
| break; |
| } |
| tcg_temp_free(z); |
| } |
| |
| static void move_to_cr(DisasContext *ctx, TCGv val, int cr) |
| { |
| TCGv z; |
| if (cr >= 8 && !is_privileged(ctx, 0)) { |
| /* Some control registers can only be written in privileged mode. */ |
| qemu_log_mask(LOG_GUEST_ERROR, |
| "disallow control register write %s", rx_crname(cr)); |
| return; |
| } |
| z = tcg_const_i32(0); |
| switch (cr) { |
| case 0: /* PSW */ |
| gen_helper_set_psw(cpu_env, val); |
| break; |
| /* case 1: to PC not supported */ |
| case 2: /* USP */ |
| tcg_gen_mov_i32(cpu_usp, val); |
| tcg_gen_movcond_i32(TCG_COND_NE, cpu_sp, |
| cpu_psw_u, z, cpu_usp, cpu_sp); |
| break; |
| case 3: /* FPSW */ |
| gen_helper_set_fpsw(cpu_env, val); |
| break; |
| case 8: /* BPSW */ |
| tcg_gen_mov_i32(cpu_bpsw, val); |
| break; |
| case 9: /* BPC */ |
| tcg_gen_mov_i32(cpu_bpc, val); |
| break; |
| case 10: /* ISP */ |
| tcg_gen_mov_i32(cpu_isp, val); |
| /* if PSW.U is 0, copy isp to r0 */ |
| tcg_gen_movcond_i32(TCG_COND_EQ, cpu_sp, |
| cpu_psw_u, z, cpu_isp, cpu_sp); |
| break; |
| case 11: /* FINTV */ |
| tcg_gen_mov_i32(cpu_fintv, val); |
| break; |
| case 12: /* INTB */ |
| tcg_gen_mov_i32(cpu_intb, val); |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, |
| "Unimplement control register %d", cr); |
| break; |
| } |
| tcg_temp_free(z); |
| } |
| |
| static void push(TCGv val) |
| { |
| tcg_gen_subi_i32(cpu_sp, cpu_sp, 4); |
| rx_gen_st(MO_32, val, cpu_sp); |
| } |
| |
| static void pop(TCGv ret) |
| { |
| rx_gen_ld(MO_32, ret, cpu_sp); |
| tcg_gen_addi_i32(cpu_sp, cpu_sp, 4); |
| } |
| |
| /* mov.<bwl> rs,dsp5[rd] */ |
| static bool trans_MOV_rm(DisasContext *ctx, arg_MOV_rm *a) |
| { |
| TCGv mem; |
| mem = tcg_temp_new(); |
| tcg_gen_addi_i32(mem, cpu_regs[a->rd], a->dsp << a->sz); |
| rx_gen_st(a->sz, cpu_regs[a->rs], mem); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* mov.<bwl> dsp5[rs],rd */ |
| static bool trans_MOV_mr(DisasContext *ctx, arg_MOV_mr *a) |
| { |
| TCGv mem; |
| mem = tcg_temp_new(); |
| tcg_gen_addi_i32(mem, cpu_regs[a->rs], a->dsp << a->sz); |
| rx_gen_ld(a->sz, cpu_regs[a->rd], mem); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* mov.l #uimm4,rd */ |
| /* mov.l #uimm8,rd */ |
| /* mov.l #imm,rd */ |
| static bool trans_MOV_ir(DisasContext *ctx, arg_MOV_ir *a) |
| { |
| tcg_gen_movi_i32(cpu_regs[a->rd], a->imm); |
| return true; |
| } |
| |
| /* mov.<bwl> #uimm8,dsp[rd] */ |
| /* mov.<bwl> #imm, dsp[rd] */ |
| static bool trans_MOV_im(DisasContext *ctx, arg_MOV_im *a) |
| { |
| TCGv imm, mem; |
| imm = tcg_const_i32(a->imm); |
| mem = tcg_temp_new(); |
| tcg_gen_addi_i32(mem, cpu_regs[a->rd], a->dsp << a->sz); |
| rx_gen_st(a->sz, imm, mem); |
| tcg_temp_free(imm); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* mov.<bwl> [ri,rb],rd */ |
| static bool trans_MOV_ar(DisasContext *ctx, arg_MOV_ar *a) |
| { |
| TCGv mem; |
| mem = tcg_temp_new(); |
| rx_gen_regindex(ctx, mem, a->sz, a->ri, a->rb); |
| rx_gen_ld(a->sz, cpu_regs[a->rd], mem); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* mov.<bwl> rd,[ri,rb] */ |
| static bool trans_MOV_ra(DisasContext *ctx, arg_MOV_ra *a) |
| { |
| TCGv mem; |
| mem = tcg_temp_new(); |
| rx_gen_regindex(ctx, mem, a->sz, a->ri, a->rb); |
| rx_gen_st(a->sz, cpu_regs[a->rs], mem); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* mov.<bwl> dsp[rs],dsp[rd] */ |
| /* mov.<bwl> rs,dsp[rd] */ |
| /* mov.<bwl> dsp[rs],rd */ |
| /* mov.<bwl> rs,rd */ |
| static bool trans_MOV_mm(DisasContext *ctx, arg_MOV_mm *a) |
| { |
| static void (* const mov[])(TCGv ret, TCGv arg) = { |
| tcg_gen_ext8s_i32, tcg_gen_ext16s_i32, tcg_gen_mov_i32, |
| }; |
| TCGv tmp, mem, addr; |
| if (a->lds == 3 && a->ldd == 3) { |
| /* mov.<bwl> rs,rd */ |
| mov[a->sz](cpu_regs[a->rd], cpu_regs[a->rs]); |
| return true; |
| } |
| |
| mem = tcg_temp_new(); |
| if (a->lds == 3) { |
| /* mov.<bwl> rs,dsp[rd] */ |
| addr = rx_index_addr(ctx, mem, a->ldd, a->sz, a->rs); |
| rx_gen_st(a->sz, cpu_regs[a->rd], addr); |
| } else if (a->ldd == 3) { |
| /* mov.<bwl> dsp[rs],rd */ |
| addr = rx_index_addr(ctx, mem, a->lds, a->sz, a->rs); |
| rx_gen_ld(a->sz, cpu_regs[a->rd], addr); |
| } else { |
| /* mov.<bwl> dsp[rs],dsp[rd] */ |
| tmp = tcg_temp_new(); |
| addr = rx_index_addr(ctx, mem, a->lds, a->sz, a->rs); |
| rx_gen_ld(a->sz, tmp, addr); |
| addr = rx_index_addr(ctx, mem, a->ldd, a->sz, a->rd); |
| rx_gen_st(a->sz, tmp, addr); |
| tcg_temp_free(tmp); |
| } |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* mov.<bwl> rs,[rd+] */ |
| /* mov.<bwl> rs,[-rd] */ |
| static bool trans_MOV_rp(DisasContext *ctx, arg_MOV_rp *a) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| tcg_gen_mov_i32(val, cpu_regs[a->rs]); |
| if (a->ad == 1) { |
| tcg_gen_subi_i32(cpu_regs[a->rd], cpu_regs[a->rd], 1 << a->sz); |
| } |
| rx_gen_st(a->sz, val, cpu_regs[a->rd]); |
| if (a->ad == 0) { |
| tcg_gen_addi_i32(cpu_regs[a->rd], cpu_regs[a->rd], 1 << a->sz); |
| } |
| tcg_temp_free(val); |
| return true; |
| } |
| |
| /* mov.<bwl> [rd+],rs */ |
| /* mov.<bwl> [-rd],rs */ |
| static bool trans_MOV_pr(DisasContext *ctx, arg_MOV_pr *a) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| if (a->ad == 1) { |
| tcg_gen_subi_i32(cpu_regs[a->rd], cpu_regs[a->rd], 1 << a->sz); |
| } |
| rx_gen_ld(a->sz, val, cpu_regs[a->rd]); |
| if (a->ad == 0) { |
| tcg_gen_addi_i32(cpu_regs[a->rd], cpu_regs[a->rd], 1 << a->sz); |
| } |
| tcg_gen_mov_i32(cpu_regs[a->rs], val); |
| tcg_temp_free(val); |
| return true; |
| } |
| |
| /* movu.<bw> dsp5[rs],rd */ |
| /* movu.<bw> dsp[rs],rd */ |
| static bool trans_MOVU_mr(DisasContext *ctx, arg_MOVU_mr *a) |
| { |
| TCGv mem; |
| mem = tcg_temp_new(); |
| tcg_gen_addi_i32(mem, cpu_regs[a->rs], a->dsp << a->sz); |
| rx_gen_ldu(a->sz, cpu_regs[a->rd], mem); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* movu.<bw> rs,rd */ |
| static bool trans_MOVU_rr(DisasContext *ctx, arg_MOVU_rr *a) |
| { |
| static void (* const ext[])(TCGv ret, TCGv arg) = { |
| tcg_gen_ext8u_i32, tcg_gen_ext16u_i32, |
| }; |
| ext[a->sz](cpu_regs[a->rd], cpu_regs[a->rs]); |
| return true; |
| } |
| |
| /* movu.<bw> [ri,rb],rd */ |
| static bool trans_MOVU_ar(DisasContext *ctx, arg_MOVU_ar *a) |
| { |
| TCGv mem; |
| mem = tcg_temp_new(); |
| rx_gen_regindex(ctx, mem, a->sz, a->ri, a->rb); |
| rx_gen_ldu(a->sz, cpu_regs[a->rd], mem); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* movu.<bw> [rd+],rs */ |
| /* mov.<bw> [-rd],rs */ |
| static bool trans_MOVU_pr(DisasContext *ctx, arg_MOVU_pr *a) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| if (a->ad == 1) { |
| tcg_gen_subi_i32(cpu_regs[a->rd], cpu_regs[a->rd], 1 << a->sz); |
| } |
| rx_gen_ldu(a->sz, val, cpu_regs[a->rd]); |
| if (a->ad == 0) { |
| tcg_gen_addi_i32(cpu_regs[a->rd], cpu_regs[a->rd], 1 << a->sz); |
| } |
| tcg_gen_mov_i32(cpu_regs[a->rs], val); |
| tcg_temp_free(val); |
| return true; |
| } |
| |
| |
| /* pop rd */ |
| static bool trans_POP(DisasContext *ctx, arg_POP *a) |
| { |
| /* mov.l [r0+], rd */ |
| arg_MOV_rp mov_a; |
| mov_a.rd = 0; |
| mov_a.rs = a->rd; |
| mov_a.ad = 0; |
| mov_a.sz = MO_32; |
| trans_MOV_pr(ctx, &mov_a); |
| return true; |
| } |
| |
| /* popc cr */ |
| static bool trans_POPC(DisasContext *ctx, arg_POPC *a) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| pop(val); |
| move_to_cr(ctx, val, a->cr); |
| if (a->cr == 0 && is_privileged(ctx, 0)) { |
| /* PSW.I may be updated here. exit TB. */ |
| ctx->base.is_jmp = DISAS_UPDATE; |
| } |
| tcg_temp_free(val); |
| return true; |
| } |
| |
| /* popm rd-rd2 */ |
| static bool trans_POPM(DisasContext *ctx, arg_POPM *a) |
| { |
| int r; |
| if (a->rd == 0 || a->rd >= a->rd2) { |
| qemu_log_mask(LOG_GUEST_ERROR, |
| "Invalid register ranges r%d-r%d", a->rd, a->rd2); |
| } |
| r = a->rd; |
| while (r <= a->rd2 && r < 16) { |
| pop(cpu_regs[r++]); |
| } |
| return true; |
| } |
| |
| |
| /* push.<bwl> rs */ |
| static bool trans_PUSH_r(DisasContext *ctx, arg_PUSH_r *a) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| tcg_gen_mov_i32(val, cpu_regs[a->rs]); |
| tcg_gen_subi_i32(cpu_sp, cpu_sp, 4); |
| rx_gen_st(a->sz, val, cpu_sp); |
| tcg_temp_free(val); |
| return true; |
| } |
| |
| /* push.<bwl> dsp[rs] */ |
| static bool trans_PUSH_m(DisasContext *ctx, arg_PUSH_m *a) |
| { |
| TCGv mem, val, addr; |
| mem = tcg_temp_new(); |
| val = tcg_temp_new(); |
| addr = rx_index_addr(ctx, mem, a->ld, a->sz, a->rs); |
| rx_gen_ld(a->sz, val, addr); |
| tcg_gen_subi_i32(cpu_sp, cpu_sp, 4); |
| rx_gen_st(a->sz, val, cpu_sp); |
| tcg_temp_free(mem); |
| tcg_temp_free(val); |
| return true; |
| } |
| |
| /* pushc rx */ |
| static bool trans_PUSHC(DisasContext *ctx, arg_PUSHC *a) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| move_from_cr(val, a->cr, ctx->pc); |
| push(val); |
| tcg_temp_free(val); |
| return true; |
| } |
| |
| /* pushm rs-rs2 */ |
| static bool trans_PUSHM(DisasContext *ctx, arg_PUSHM *a) |
| { |
| int r; |
| |
| if (a->rs == 0 || a->rs >= a->rs2) { |
| qemu_log_mask(LOG_GUEST_ERROR, |
| "Invalid register ranges r%d-r%d", a->rs, a->rs2); |
| } |
| r = a->rs2; |
| while (r >= a->rs && r >= 0) { |
| push(cpu_regs[r--]); |
| } |
| return true; |
| } |
| |
| /* xchg rs,rd */ |
| static bool trans_XCHG_rr(DisasContext *ctx, arg_XCHG_rr *a) |
| { |
| TCGv tmp; |
| tmp = tcg_temp_new(); |
| tcg_gen_mov_i32(tmp, cpu_regs[a->rs]); |
| tcg_gen_mov_i32(cpu_regs[a->rs], cpu_regs[a->rd]); |
| tcg_gen_mov_i32(cpu_regs[a->rd], tmp); |
| tcg_temp_free(tmp); |
| return true; |
| } |
| |
| /* xchg dsp[rs].<mi>,rd */ |
| static bool trans_XCHG_mr(DisasContext *ctx, arg_XCHG_mr *a) |
| { |
| TCGv mem, addr; |
| mem = tcg_temp_new(); |
| switch (a->mi) { |
| case 0: /* dsp[rs].b */ |
| case 1: /* dsp[rs].w */ |
| case 2: /* dsp[rs].l */ |
| addr = rx_index_addr(ctx, mem, a->ld, a->mi, a->rs); |
| break; |
| case 3: /* dsp[rs].uw */ |
| case 4: /* dsp[rs].ub */ |
| addr = rx_index_addr(ctx, mem, a->ld, 4 - a->mi, a->rs); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| tcg_gen_atomic_xchg_i32(cpu_regs[a->rd], addr, cpu_regs[a->rd], |
| 0, mi_to_mop(a->mi)); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| static inline void stcond(TCGCond cond, int rd, int imm) |
| { |
| TCGv z; |
| TCGv _imm; |
| z = tcg_const_i32(0); |
| _imm = tcg_const_i32(imm); |
| tcg_gen_movcond_i32(cond, cpu_regs[rd], cpu_psw_z, z, |
| _imm, cpu_regs[rd]); |
| tcg_temp_free(z); |
| tcg_temp_free(_imm); |
| } |
| |
| /* stz #imm,rd */ |
| static bool trans_STZ(DisasContext *ctx, arg_STZ *a) |
| { |
| stcond(TCG_COND_EQ, a->rd, a->imm); |
| return true; |
| } |
| |
| /* stnz #imm,rd */ |
| static bool trans_STNZ(DisasContext *ctx, arg_STNZ *a) |
| { |
| stcond(TCG_COND_NE, a->rd, a->imm); |
| return true; |
| } |
| |
| /* sccnd.<bwl> rd */ |
| /* sccnd.<bwl> dsp:[rd] */ |
| static bool trans_SCCnd(DisasContext *ctx, arg_SCCnd *a) |
| { |
| DisasCompare dc; |
| TCGv val, mem, addr; |
| dc.temp = tcg_temp_new(); |
| psw_cond(&dc, a->cd); |
| if (a->ld < 3) { |
| val = tcg_temp_new(); |
| mem = tcg_temp_new(); |
| tcg_gen_setcondi_i32(dc.cond, val, dc.value, 0); |
| addr = rx_index_addr(ctx, mem, a->sz, a->ld, a->rd); |
| rx_gen_st(a->sz, val, addr); |
| tcg_temp_free(val); |
| tcg_temp_free(mem); |
| } else { |
| tcg_gen_setcondi_i32(dc.cond, cpu_regs[a->rd], dc.value, 0); |
| } |
| tcg_temp_free(dc.temp); |
| return true; |
| } |
| |
| /* rtsd #imm */ |
| static bool trans_RTSD_i(DisasContext *ctx, arg_RTSD_i *a) |
| { |
| tcg_gen_addi_i32(cpu_sp, cpu_sp, a->imm << 2); |
| pop(cpu_pc); |
| ctx->base.is_jmp = DISAS_JUMP; |
| return true; |
| } |
| |
| /* rtsd #imm, rd-rd2 */ |
| static bool trans_RTSD_irr(DisasContext *ctx, arg_RTSD_irr *a) |
| { |
| int dst; |
| int adj; |
| |
| if (a->rd2 >= a->rd) { |
| adj = a->imm - (a->rd2 - a->rd + 1); |
| } else { |
| adj = a->imm - (15 - a->rd + 1); |
| } |
| |
| tcg_gen_addi_i32(cpu_sp, cpu_sp, adj << 2); |
| dst = a->rd; |
| while (dst <= a->rd2 && dst < 16) { |
| pop(cpu_regs[dst++]); |
| } |
| pop(cpu_pc); |
| ctx->base.is_jmp = DISAS_JUMP; |
| return true; |
| } |
| |
| typedef void (*op2fn)(TCGv ret, TCGv arg1); |
| typedef void (*op3fn)(TCGv ret, TCGv arg1, TCGv arg2); |
| |
| static inline void rx_gen_op_rr(op2fn opr, int dst, int src) |
| { |
| opr(cpu_regs[dst], cpu_regs[src]); |
| } |
| |
| static inline void rx_gen_op_rrr(op3fn opr, int dst, int src, int src2) |
| { |
| opr(cpu_regs[dst], cpu_regs[src], cpu_regs[src2]); |
| } |
| |
| static inline void rx_gen_op_irr(op3fn opr, int dst, int src, uint32_t src2) |
| { |
| TCGv imm = tcg_const_i32(src2); |
| opr(cpu_regs[dst], cpu_regs[src], imm); |
| tcg_temp_free(imm); |
| } |
| |
| static inline void rx_gen_op_mr(op3fn opr, DisasContext *ctx, |
| int dst, int src, int ld, int mi) |
| { |
| TCGv val, mem; |
| mem = tcg_temp_new(); |
| val = rx_load_source(ctx, mem, ld, mi, src); |
| opr(cpu_regs[dst], cpu_regs[dst], val); |
| tcg_temp_free(mem); |
| } |
| |
| static void rx_and(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| tcg_gen_and_i32(cpu_psw_s, arg1, arg2); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_s); |
| tcg_gen_mov_i32(ret, cpu_psw_s); |
| } |
| |
| /* and #uimm:4, rd */ |
| /* and #imm, rd */ |
| static bool trans_AND_ir(DisasContext *ctx, arg_AND_ir *a) |
| { |
| rx_gen_op_irr(rx_and, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* and dsp[rs], rd */ |
| /* and rs,rd */ |
| static bool trans_AND_mr(DisasContext *ctx, arg_AND_mr *a) |
| { |
| rx_gen_op_mr(rx_and, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* and rs,rs2,rd */ |
| static bool trans_AND_rrr(DisasContext *ctx, arg_AND_rrr *a) |
| { |
| rx_gen_op_rrr(rx_and, a->rd, a->rs, a->rs2); |
| return true; |
| } |
| |
| static void rx_or(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| tcg_gen_or_i32(cpu_psw_s, arg1, arg2); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_s); |
| tcg_gen_mov_i32(ret, cpu_psw_s); |
| } |
| |
| /* or #uimm:4, rd */ |
| /* or #imm, rd */ |
| static bool trans_OR_ir(DisasContext *ctx, arg_OR_ir *a) |
| { |
| rx_gen_op_irr(rx_or, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* or dsp[rs], rd */ |
| /* or rs,rd */ |
| static bool trans_OR_mr(DisasContext *ctx, arg_OR_mr *a) |
| { |
| rx_gen_op_mr(rx_or, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* or rs,rs2,rd */ |
| static bool trans_OR_rrr(DisasContext *ctx, arg_OR_rrr *a) |
| { |
| rx_gen_op_rrr(rx_or, a->rd, a->rs, a->rs2); |
| return true; |
| } |
| |
| static void rx_xor(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| tcg_gen_xor_i32(cpu_psw_s, arg1, arg2); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_s); |
| tcg_gen_mov_i32(ret, cpu_psw_s); |
| } |
| |
| /* xor #imm, rd */ |
| static bool trans_XOR_ir(DisasContext *ctx, arg_XOR_ir *a) |
| { |
| rx_gen_op_irr(rx_xor, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* xor dsp[rs], rd */ |
| /* xor rs,rd */ |
| static bool trans_XOR_mr(DisasContext *ctx, arg_XOR_mr *a) |
| { |
| rx_gen_op_mr(rx_xor, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| static void rx_tst(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| tcg_gen_and_i32(cpu_psw_s, arg1, arg2); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_s); |
| } |
| |
| /* tst #imm, rd */ |
| static bool trans_TST_ir(DisasContext *ctx, arg_TST_ir *a) |
| { |
| rx_gen_op_irr(rx_tst, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* tst dsp[rs], rd */ |
| /* tst rs, rd */ |
| static bool trans_TST_mr(DisasContext *ctx, arg_TST_mr *a) |
| { |
| rx_gen_op_mr(rx_tst, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| static void rx_not(TCGv ret, TCGv arg1) |
| { |
| tcg_gen_not_i32(ret, arg1); |
| tcg_gen_mov_i32(cpu_psw_z, ret); |
| tcg_gen_mov_i32(cpu_psw_s, ret); |
| } |
| |
| /* not rd */ |
| /* not rs, rd */ |
| static bool trans_NOT_rr(DisasContext *ctx, arg_NOT_rr *a) |
| { |
| rx_gen_op_rr(rx_not, a->rd, a->rs); |
| return true; |
| } |
| |
| static void rx_neg(TCGv ret, TCGv arg1) |
| { |
| tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_psw_o, arg1, 0x80000000); |
| tcg_gen_neg_i32(ret, arg1); |
| tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_psw_c, ret, 0); |
| tcg_gen_mov_i32(cpu_psw_z, ret); |
| tcg_gen_mov_i32(cpu_psw_s, ret); |
| } |
| |
| |
| /* neg rd */ |
| /* neg rs, rd */ |
| static bool trans_NEG_rr(DisasContext *ctx, arg_NEG_rr *a) |
| { |
| rx_gen_op_rr(rx_neg, a->rd, a->rs); |
| return true; |
| } |
| |
| /* ret = arg1 + arg2 + psw_c */ |
| static void rx_adc(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| TCGv z; |
| z = tcg_const_i32(0); |
| tcg_gen_add2_i32(cpu_psw_s, cpu_psw_c, arg1, z, cpu_psw_c, z); |
| tcg_gen_add2_i32(cpu_psw_s, cpu_psw_c, cpu_psw_s, cpu_psw_c, arg2, z); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_s); |
| tcg_gen_xor_i32(cpu_psw_o, cpu_psw_s, arg1); |
| tcg_gen_xor_i32(z, arg1, arg2); |
| tcg_gen_andc_i32(cpu_psw_o, cpu_psw_o, z); |
| tcg_gen_mov_i32(ret, cpu_psw_s); |
| tcg_temp_free(z); |
| } |
| |
| /* adc #imm, rd */ |
| static bool trans_ADC_ir(DisasContext *ctx, arg_ADC_ir *a) |
| { |
| rx_gen_op_irr(rx_adc, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* adc rs, rd */ |
| static bool trans_ADC_rr(DisasContext *ctx, arg_ADC_rr *a) |
| { |
| rx_gen_op_rrr(rx_adc, a->rd, a->rd, a->rs); |
| return true; |
| } |
| |
| /* adc dsp[rs], rd */ |
| static bool trans_ADC_mr(DisasContext *ctx, arg_ADC_mr *a) |
| { |
| /* mi only 2 */ |
| if (a->mi != 2) { |
| return false; |
| } |
| rx_gen_op_mr(rx_adc, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* ret = arg1 + arg2 */ |
| static void rx_add(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| TCGv z; |
| z = tcg_const_i32(0); |
| tcg_gen_add2_i32(cpu_psw_s, cpu_psw_c, arg1, z, arg2, z); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_s); |
| tcg_gen_xor_i32(cpu_psw_o, cpu_psw_s, arg1); |
| tcg_gen_xor_i32(z, arg1, arg2); |
| tcg_gen_andc_i32(cpu_psw_o, cpu_psw_o, z); |
| tcg_gen_mov_i32(ret, cpu_psw_s); |
| tcg_temp_free(z); |
| } |
| |
| /* add #uimm4, rd */ |
| /* add #imm, rs, rd */ |
| static bool trans_ADD_irr(DisasContext *ctx, arg_ADD_irr *a) |
| { |
| rx_gen_op_irr(rx_add, a->rd, a->rs2, a->imm); |
| return true; |
| } |
| |
| /* add rs, rd */ |
| /* add dsp[rs], rd */ |
| static bool trans_ADD_mr(DisasContext *ctx, arg_ADD_mr *a) |
| { |
| rx_gen_op_mr(rx_add, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* add rs, rs2, rd */ |
| static bool trans_ADD_rrr(DisasContext *ctx, arg_ADD_rrr *a) |
| { |
| rx_gen_op_rrr(rx_add, a->rd, a->rs, a->rs2); |
| return true; |
| } |
| |
| /* ret = arg1 - arg2 */ |
| static void rx_sub(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| TCGv temp; |
| tcg_gen_sub_i32(cpu_psw_s, arg1, arg2); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_s); |
| tcg_gen_setcond_i32(TCG_COND_GEU, cpu_psw_c, arg1, arg2); |
| tcg_gen_xor_i32(cpu_psw_o, cpu_psw_s, arg1); |
| temp = tcg_temp_new_i32(); |
| tcg_gen_xor_i32(temp, arg1, arg2); |
| tcg_gen_and_i32(cpu_psw_o, cpu_psw_o, temp); |
| tcg_temp_free_i32(temp); |
| /* CMP not required return */ |
| if (ret) { |
| tcg_gen_mov_i32(ret, cpu_psw_s); |
| } |
| } |
| static void rx_cmp(TCGv dummy, TCGv arg1, TCGv arg2) |
| { |
| rx_sub(NULL, arg1, arg2); |
| } |
| /* ret = arg1 - arg2 - !psw_c */ |
| /* -> ret = arg1 + ~arg2 + psw_c */ |
| static void rx_sbb(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| TCGv temp; |
| temp = tcg_temp_new(); |
| tcg_gen_not_i32(temp, arg2); |
| rx_adc(ret, arg1, temp); |
| tcg_temp_free(temp); |
| } |
| |
| /* cmp #imm4, rs2 */ |
| /* cmp #imm8, rs2 */ |
| /* cmp #imm, rs2 */ |
| static bool trans_CMP_ir(DisasContext *ctx, arg_CMP_ir *a) |
| { |
| rx_gen_op_irr(rx_cmp, 0, a->rs2, a->imm); |
| return true; |
| } |
| |
| /* cmp rs, rs2 */ |
| /* cmp dsp[rs], rs2 */ |
| static bool trans_CMP_mr(DisasContext *ctx, arg_CMP_mr *a) |
| { |
| rx_gen_op_mr(rx_cmp, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* sub #imm4, rd */ |
| static bool trans_SUB_ir(DisasContext *ctx, arg_SUB_ir *a) |
| { |
| rx_gen_op_irr(rx_sub, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* sub rs, rd */ |
| /* sub dsp[rs], rd */ |
| static bool trans_SUB_mr(DisasContext *ctx, arg_SUB_mr *a) |
| { |
| rx_gen_op_mr(rx_sub, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* sub rs2, rs, rd */ |
| static bool trans_SUB_rrr(DisasContext *ctx, arg_SUB_rrr *a) |
| { |
| rx_gen_op_rrr(rx_sub, a->rd, a->rs2, a->rs); |
| return true; |
| } |
| |
| /* sbb rs, rd */ |
| static bool trans_SBB_rr(DisasContext *ctx, arg_SBB_rr *a) |
| { |
| rx_gen_op_rrr(rx_sbb, a->rd, a->rd, a->rs); |
| return true; |
| } |
| |
| /* sbb dsp[rs], rd */ |
| static bool trans_SBB_mr(DisasContext *ctx, arg_SBB_mr *a) |
| { |
| /* mi only 2 */ |
| if (a->mi != 2) { |
| return false; |
| } |
| rx_gen_op_mr(rx_sbb, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| static void rx_abs(TCGv ret, TCGv arg1) |
| { |
| TCGv neg; |
| TCGv zero; |
| neg = tcg_temp_new(); |
| zero = tcg_const_i32(0); |
| tcg_gen_neg_i32(neg, arg1); |
| tcg_gen_movcond_i32(TCG_COND_LT, ret, arg1, zero, neg, arg1); |
| tcg_temp_free(neg); |
| tcg_temp_free(zero); |
| } |
| |
| /* abs rd */ |
| /* abs rs, rd */ |
| static bool trans_ABS_rr(DisasContext *ctx, arg_ABS_rr *a) |
| { |
| rx_gen_op_rr(rx_abs, a->rd, a->rs); |
| return true; |
| } |
| |
| /* max #imm, rd */ |
| static bool trans_MAX_ir(DisasContext *ctx, arg_MAX_ir *a) |
| { |
| rx_gen_op_irr(tcg_gen_smax_i32, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* max rs, rd */ |
| /* max dsp[rs], rd */ |
| static bool trans_MAX_mr(DisasContext *ctx, arg_MAX_mr *a) |
| { |
| rx_gen_op_mr(tcg_gen_smax_i32, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* min #imm, rd */ |
| static bool trans_MIN_ir(DisasContext *ctx, arg_MIN_ir *a) |
| { |
| rx_gen_op_irr(tcg_gen_smin_i32, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* min rs, rd */ |
| /* min dsp[rs], rd */ |
| static bool trans_MIN_mr(DisasContext *ctx, arg_MIN_mr *a) |
| { |
| rx_gen_op_mr(tcg_gen_smin_i32, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* mul #uimm4, rd */ |
| /* mul #imm, rd */ |
| static bool trans_MUL_ir(DisasContext *ctx, arg_MUL_ir *a) |
| { |
| rx_gen_op_irr(tcg_gen_mul_i32, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* mul rs, rd */ |
| /* mul dsp[rs], rd */ |
| static bool trans_MUL_mr(DisasContext *ctx, arg_MUL_mr *a) |
| { |
| rx_gen_op_mr(tcg_gen_mul_i32, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* mul rs, rs2, rd */ |
| static bool trans_MUL_rrr(DisasContext *ctx, arg_MUL_rrr *a) |
| { |
| rx_gen_op_rrr(tcg_gen_mul_i32, a->rd, a->rs, a->rs2); |
| return true; |
| } |
| |
| /* emul #imm, rd */ |
| static bool trans_EMUL_ir(DisasContext *ctx, arg_EMUL_ir *a) |
| { |
| TCGv imm = tcg_const_i32(a->imm); |
| if (a->rd > 14) { |
| qemu_log_mask(LOG_GUEST_ERROR, "rd too large %d", a->rd); |
| } |
| tcg_gen_muls2_i32(cpu_regs[a->rd], cpu_regs[(a->rd + 1) & 15], |
| cpu_regs[a->rd], imm); |
| tcg_temp_free(imm); |
| return true; |
| } |
| |
| /* emul rs, rd */ |
| /* emul dsp[rs], rd */ |
| static bool trans_EMUL_mr(DisasContext *ctx, arg_EMUL_mr *a) |
| { |
| TCGv val, mem; |
| if (a->rd > 14) { |
| qemu_log_mask(LOG_GUEST_ERROR, "rd too large %d", a->rd); |
| } |
| mem = tcg_temp_new(); |
| val = rx_load_source(ctx, mem, a->ld, a->mi, a->rs); |
| tcg_gen_muls2_i32(cpu_regs[a->rd], cpu_regs[(a->rd + 1) & 15], |
| cpu_regs[a->rd], val); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* emulu #imm, rd */ |
| static bool trans_EMULU_ir(DisasContext *ctx, arg_EMULU_ir *a) |
| { |
| TCGv imm = tcg_const_i32(a->imm); |
| if (a->rd > 14) { |
| qemu_log_mask(LOG_GUEST_ERROR, "rd too large %d", a->rd); |
| } |
| tcg_gen_mulu2_i32(cpu_regs[a->rd], cpu_regs[(a->rd + 1) & 15], |
| cpu_regs[a->rd], imm); |
| tcg_temp_free(imm); |
| return true; |
| } |
| |
| /* emulu rs, rd */ |
| /* emulu dsp[rs], rd */ |
| static bool trans_EMULU_mr(DisasContext *ctx, arg_EMULU_mr *a) |
| { |
| TCGv val, mem; |
| if (a->rd > 14) { |
| qemu_log_mask(LOG_GUEST_ERROR, "rd too large %d", a->rd); |
| } |
| mem = tcg_temp_new(); |
| val = rx_load_source(ctx, mem, a->ld, a->mi, a->rs); |
| tcg_gen_mulu2_i32(cpu_regs[a->rd], cpu_regs[(a->rd + 1) & 15], |
| cpu_regs[a->rd], val); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| static void rx_div(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| gen_helper_div(ret, cpu_env, arg1, arg2); |
| } |
| |
| static void rx_divu(TCGv ret, TCGv arg1, TCGv arg2) |
| { |
| gen_helper_divu(ret, cpu_env, arg1, arg2); |
| } |
| |
| /* div #imm, rd */ |
| static bool trans_DIV_ir(DisasContext *ctx, arg_DIV_ir *a) |
| { |
| rx_gen_op_irr(rx_div, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* div rs, rd */ |
| /* div dsp[rs], rd */ |
| static bool trans_DIV_mr(DisasContext *ctx, arg_DIV_mr *a) |
| { |
| rx_gen_op_mr(rx_div, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| /* divu #imm, rd */ |
| static bool trans_DIVU_ir(DisasContext *ctx, arg_DIVU_ir *a) |
| { |
| rx_gen_op_irr(rx_divu, a->rd, a->rd, a->imm); |
| return true; |
| } |
| |
| /* divu rs, rd */ |
| /* divu dsp[rs], rd */ |
| static bool trans_DIVU_mr(DisasContext *ctx, arg_DIVU_mr *a) |
| { |
| rx_gen_op_mr(rx_divu, ctx, a->rd, a->rs, a->ld, a->mi); |
| return true; |
| } |
| |
| |
| /* shll #imm:5, rd */ |
| /* shll #imm:5, rs2, rd */ |
| static bool trans_SHLL_irr(DisasContext *ctx, arg_SHLL_irr *a) |
| { |
| TCGv tmp; |
| tmp = tcg_temp_new(); |
| if (a->imm) { |
| tcg_gen_sari_i32(cpu_psw_c, cpu_regs[a->rs2], 32 - a->imm); |
| tcg_gen_shli_i32(cpu_regs[a->rd], cpu_regs[a->rs2], a->imm); |
| tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_psw_o, cpu_psw_c, 0); |
| tcg_gen_setcondi_i32(TCG_COND_EQ, tmp, cpu_psw_c, 0xffffffff); |
| tcg_gen_or_i32(cpu_psw_o, cpu_psw_o, tmp); |
| tcg_gen_setcondi_i32(TCG_COND_NE, cpu_psw_c, cpu_psw_c, 0); |
| } else { |
| tcg_gen_mov_i32(cpu_regs[a->rd], cpu_regs[a->rs2]); |
| tcg_gen_movi_i32(cpu_psw_c, 0); |
| tcg_gen_movi_i32(cpu_psw_o, 0); |
| } |
| tcg_gen_mov_i32(cpu_psw_z, cpu_regs[a->rd]); |
| tcg_gen_mov_i32(cpu_psw_s, cpu_regs[a->rd]); |
| return true; |
| } |
| |
| /* shll rs, rd */ |
| static bool trans_SHLL_rr(DisasContext *ctx, arg_SHLL_rr *a) |
| { |
| TCGLabel *noshift, *done; |
| TCGv count, tmp; |
| |
| noshift = gen_new_label(); |
| done = gen_new_label(); |
| /* if (cpu_regs[a->rs]) { */ |
| tcg_gen_brcondi_i32(TCG_COND_EQ, cpu_regs[a->rs], 0, noshift); |
| count = tcg_const_i32(32); |
| tmp = tcg_temp_new(); |
| tcg_gen_andi_i32(tmp, cpu_regs[a->rs], 31); |
| tcg_gen_sub_i32(count, count, tmp); |
| tcg_gen_sar_i32(cpu_psw_c, cpu_regs[a->rd], count); |
| tcg_gen_shl_i32(cpu_regs[a->rd], cpu_regs[a->rd], tmp); |
| tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_psw_o, cpu_psw_c, 0); |
| tcg_gen_setcondi_i32(TCG_COND_EQ, tmp, cpu_psw_c, 0xffffffff); |
| tcg_gen_or_i32(cpu_psw_o, cpu_psw_o, tmp); |
| tcg_gen_setcondi_i32(TCG_COND_NE, cpu_psw_c, cpu_psw_c, 0); |
| tcg_gen_br(done); |
| /* } else { */ |
| gen_set_label(noshift); |
| tcg_gen_movi_i32(cpu_psw_c, 0); |
| tcg_gen_movi_i32(cpu_psw_o, 0); |
| /* } */ |
| gen_set_label(done); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_regs[a->rd]); |
| tcg_gen_mov_i32(cpu_psw_s, cpu_regs[a->rd]); |
| tcg_temp_free(count); |
| tcg_temp_free(tmp); |
| return true; |
| } |
| |
| static inline void shiftr_imm(uint32_t rd, uint32_t rs, uint32_t imm, |
| unsigned int alith) |
| { |
| static void (* const gen_sXri[])(TCGv ret, TCGv arg1, int arg2) = { |
| tcg_gen_shri_i32, tcg_gen_sari_i32, |
| }; |
| tcg_debug_assert(alith < 2); |
| if (imm) { |
| gen_sXri[alith](cpu_regs[rd], cpu_regs[rs], imm - 1); |
| tcg_gen_andi_i32(cpu_psw_c, cpu_regs[rd], 0x00000001); |
| gen_sXri[alith](cpu_regs[rd], cpu_regs[rd], 1); |
| } else { |
| tcg_gen_mov_i32(cpu_regs[rd], cpu_regs[rs]); |
| tcg_gen_movi_i32(cpu_psw_c, 0); |
| } |
| tcg_gen_movi_i32(cpu_psw_o, 0); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_regs[rd]); |
| tcg_gen_mov_i32(cpu_psw_s, cpu_regs[rd]); |
| } |
| |
| static inline void shiftr_reg(uint32_t rd, uint32_t rs, unsigned int alith) |
| { |
| TCGLabel *noshift, *done; |
| TCGv count; |
| static void (* const gen_sXri[])(TCGv ret, TCGv arg1, int arg2) = { |
| tcg_gen_shri_i32, tcg_gen_sari_i32, |
| }; |
| static void (* const gen_sXr[])(TCGv ret, TCGv arg1, TCGv arg2) = { |
| tcg_gen_shr_i32, tcg_gen_sar_i32, |
| }; |
| tcg_debug_assert(alith < 2); |
| noshift = gen_new_label(); |
| done = gen_new_label(); |
| count = tcg_temp_new(); |
| /* if (cpu_regs[rs]) { */ |
| tcg_gen_brcondi_i32(TCG_COND_EQ, cpu_regs[rs], 0, noshift); |
| tcg_gen_andi_i32(count, cpu_regs[rs], 31); |
| tcg_gen_subi_i32(count, count, 1); |
| gen_sXr[alith](cpu_regs[rd], cpu_regs[rd], count); |
| tcg_gen_andi_i32(cpu_psw_c, cpu_regs[rd], 0x00000001); |
| gen_sXri[alith](cpu_regs[rd], cpu_regs[rd], 1); |
| tcg_gen_br(done); |
| /* } else { */ |
| gen_set_label(noshift); |
| tcg_gen_movi_i32(cpu_psw_c, 0); |
| /* } */ |
| gen_set_label(done); |
| tcg_gen_movi_i32(cpu_psw_o, 0); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_regs[rd]); |
| tcg_gen_mov_i32(cpu_psw_s, cpu_regs[rd]); |
| tcg_temp_free(count); |
| } |
| |
| /* shar #imm:5, rd */ |
| /* shar #imm:5, rs2, rd */ |
| static bool trans_SHAR_irr(DisasContext *ctx, arg_SHAR_irr *a) |
| { |
| shiftr_imm(a->rd, a->rs2, a->imm, 1); |
| return true; |
| } |
| |
| /* shar rs, rd */ |
| static bool trans_SHAR_rr(DisasContext *ctx, arg_SHAR_rr *a) |
| { |
| shiftr_reg(a->rd, a->rs, 1); |
| return true; |
| } |
| |
| /* shlr #imm:5, rd */ |
| /* shlr #imm:5, rs2, rd */ |
| static bool trans_SHLR_irr(DisasContext *ctx, arg_SHLR_irr *a) |
| { |
| shiftr_imm(a->rd, a->rs2, a->imm, 0); |
| return true; |
| } |
| |
| /* shlr rs, rd */ |
| static bool trans_SHLR_rr(DisasContext *ctx, arg_SHLR_rr *a) |
| { |
| shiftr_reg(a->rd, a->rs, 0); |
| return true; |
| } |
| |
| /* rolc rd */ |
| static bool trans_ROLC(DisasContext *ctx, arg_ROLC *a) |
| { |
| TCGv tmp; |
| tmp = tcg_temp_new(); |
| tcg_gen_shri_i32(tmp, cpu_regs[a->rd], 31); |
| tcg_gen_shli_i32(cpu_regs[a->rd], cpu_regs[a->rd], 1); |
| tcg_gen_or_i32(cpu_regs[a->rd], cpu_regs[a->rd], cpu_psw_c); |
| tcg_gen_mov_i32(cpu_psw_c, tmp); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_regs[a->rd]); |
| tcg_gen_mov_i32(cpu_psw_s, cpu_regs[a->rd]); |
| tcg_temp_free(tmp); |
| return true; |
| } |
| |
| /* rorc rd */ |
| static bool trans_RORC(DisasContext *ctx, arg_RORC *a) |
| { |
| TCGv tmp; |
| tmp = tcg_temp_new(); |
| tcg_gen_andi_i32(tmp, cpu_regs[a->rd], 0x00000001); |
| tcg_gen_shri_i32(cpu_regs[a->rd], cpu_regs[a->rd], 1); |
| tcg_gen_shli_i32(cpu_psw_c, cpu_psw_c, 31); |
| tcg_gen_or_i32(cpu_regs[a->rd], cpu_regs[a->rd], cpu_psw_c); |
| tcg_gen_mov_i32(cpu_psw_c, tmp); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_regs[a->rd]); |
| tcg_gen_mov_i32(cpu_psw_s, cpu_regs[a->rd]); |
| return true; |
| } |
| |
| enum {ROTR = 0, ROTL = 1}; |
| enum {ROT_IMM = 0, ROT_REG = 1}; |
| static inline void rx_rot(int ir, int dir, int rd, int src) |
| { |
| switch (dir) { |
| case ROTL: |
| if (ir == ROT_IMM) { |
| tcg_gen_rotli_i32(cpu_regs[rd], cpu_regs[rd], src); |
| } else { |
| tcg_gen_rotl_i32(cpu_regs[rd], cpu_regs[rd], cpu_regs[src]); |
| } |
| tcg_gen_andi_i32(cpu_psw_c, cpu_regs[rd], 0x00000001); |
| break; |
| case ROTR: |
| if (ir == ROT_IMM) { |
| tcg_gen_rotri_i32(cpu_regs[rd], cpu_regs[rd], src); |
| } else { |
| tcg_gen_rotr_i32(cpu_regs[rd], cpu_regs[rd], cpu_regs[src]); |
| } |
| tcg_gen_shri_i32(cpu_psw_c, cpu_regs[rd], 31); |
| break; |
| } |
| tcg_gen_mov_i32(cpu_psw_z, cpu_regs[rd]); |
| tcg_gen_mov_i32(cpu_psw_s, cpu_regs[rd]); |
| } |
| |
| /* rotl #imm, rd */ |
| static bool trans_ROTL_ir(DisasContext *ctx, arg_ROTL_ir *a) |
| { |
| rx_rot(ROT_IMM, ROTL, a->rd, a->imm); |
| return true; |
| } |
| |
| /* rotl rs, rd */ |
| static bool trans_ROTL_rr(DisasContext *ctx, arg_ROTL_rr *a) |
| { |
| rx_rot(ROT_REG, ROTL, a->rd, a->rs); |
| return true; |
| } |
| |
| /* rotr #imm, rd */ |
| static bool trans_ROTR_ir(DisasContext *ctx, arg_ROTR_ir *a) |
| { |
| rx_rot(ROT_IMM, ROTR, a->rd, a->imm); |
| return true; |
| } |
| |
| /* rotr rs, rd */ |
| static bool trans_ROTR_rr(DisasContext *ctx, arg_ROTR_rr *a) |
| { |
| rx_rot(ROT_REG, ROTR, a->rd, a->rs); |
| return true; |
| } |
| |
| /* revl rs, rd */ |
| static bool trans_REVL(DisasContext *ctx, arg_REVL *a) |
| { |
| tcg_gen_bswap32_i32(cpu_regs[a->rd], cpu_regs[a->rs]); |
| return true; |
| } |
| |
| /* revw rs, rd */ |
| static bool trans_REVW(DisasContext *ctx, arg_REVW *a) |
| { |
| TCGv tmp; |
| tmp = tcg_temp_new(); |
| tcg_gen_andi_i32(tmp, cpu_regs[a->rs], 0x00ff00ff); |
| tcg_gen_shli_i32(tmp, tmp, 8); |
| tcg_gen_shri_i32(cpu_regs[a->rd], cpu_regs[a->rs], 8); |
| tcg_gen_andi_i32(cpu_regs[a->rd], cpu_regs[a->rd], 0x00ff00ff); |
| tcg_gen_or_i32(cpu_regs[a->rd], cpu_regs[a->rd], tmp); |
| tcg_temp_free(tmp); |
| return true; |
| } |
| |
| /* conditional branch helper */ |
| static void rx_bcnd_main(DisasContext *ctx, int cd, int dst) |
| { |
| DisasCompare dc; |
| TCGLabel *t, *done; |
| |
| switch (cd) { |
| case 0 ... 13: |
| dc.temp = tcg_temp_new(); |
| psw_cond(&dc, cd); |
| t = gen_new_label(); |
| done = gen_new_label(); |
| tcg_gen_brcondi_i32(dc.cond, dc.value, 0, t); |
| gen_goto_tb(ctx, 0, ctx->base.pc_next); |
| tcg_gen_br(done); |
| gen_set_label(t); |
| gen_goto_tb(ctx, 1, ctx->pc + dst); |
| gen_set_label(done); |
| tcg_temp_free(dc.temp); |
| break; |
| case 14: |
| /* always true case */ |
| gen_goto_tb(ctx, 0, ctx->pc + dst); |
| break; |
| case 15: |
| /* always false case */ |
| /* Nothing do */ |
| break; |
| } |
| } |
| |
| /* beq dsp:3 / bne dsp:3 */ |
| /* beq dsp:8 / bne dsp:8 */ |
| /* bc dsp:8 / bnc dsp:8 */ |
| /* bgtu dsp:8 / bleu dsp:8 */ |
| /* bpz dsp:8 / bn dsp:8 */ |
| /* bge dsp:8 / blt dsp:8 */ |
| /* bgt dsp:8 / ble dsp:8 */ |
| /* bo dsp:8 / bno dsp:8 */ |
| /* beq dsp:16 / bne dsp:16 */ |
| static bool trans_BCnd(DisasContext *ctx, arg_BCnd *a) |
| { |
| rx_bcnd_main(ctx, a->cd, a->dsp); |
| return true; |
| } |
| |
| /* bra dsp:3 */ |
| /* bra dsp:8 */ |
| /* bra dsp:16 */ |
| /* bra dsp:24 */ |
| static bool trans_BRA(DisasContext *ctx, arg_BRA *a) |
| { |
| rx_bcnd_main(ctx, 14, a->dsp); |
| return true; |
| } |
| |
| /* bra rs */ |
| static bool trans_BRA_l(DisasContext *ctx, arg_BRA_l *a) |
| { |
| tcg_gen_addi_i32(cpu_pc, cpu_regs[a->rd], ctx->pc); |
| ctx->base.is_jmp = DISAS_JUMP; |
| return true; |
| } |
| |
| static inline void rx_save_pc(DisasContext *ctx) |
| { |
| TCGv pc = tcg_const_i32(ctx->base.pc_next); |
| push(pc); |
| tcg_temp_free(pc); |
| } |
| |
| /* jmp rs */ |
| static bool trans_JMP(DisasContext *ctx, arg_JMP *a) |
| { |
| tcg_gen_mov_i32(cpu_pc, cpu_regs[a->rs]); |
| ctx->base.is_jmp = DISAS_JUMP; |
| return true; |
| } |
| |
| /* jsr rs */ |
| static bool trans_JSR(DisasContext *ctx, arg_JSR *a) |
| { |
| rx_save_pc(ctx); |
| tcg_gen_mov_i32(cpu_pc, cpu_regs[a->rs]); |
| ctx->base.is_jmp = DISAS_JUMP; |
| return true; |
| } |
| |
| /* bsr dsp:16 */ |
| /* bsr dsp:24 */ |
| static bool trans_BSR(DisasContext *ctx, arg_BSR *a) |
| { |
| rx_save_pc(ctx); |
| rx_bcnd_main(ctx, 14, a->dsp); |
| return true; |
| } |
| |
| /* bsr rs */ |
| static bool trans_BSR_l(DisasContext *ctx, arg_BSR_l *a) |
| { |
| rx_save_pc(ctx); |
| tcg_gen_addi_i32(cpu_pc, cpu_regs[a->rd], ctx->pc); |
| ctx->base.is_jmp = DISAS_JUMP; |
| return true; |
| } |
| |
| /* rts */ |
| static bool trans_RTS(DisasContext *ctx, arg_RTS *a) |
| { |
| pop(cpu_pc); |
| ctx->base.is_jmp = DISAS_JUMP; |
| return true; |
| } |
| |
| /* nop */ |
| static bool trans_NOP(DisasContext *ctx, arg_NOP *a) |
| { |
| return true; |
| } |
| |
| /* scmpu */ |
| static bool trans_SCMPU(DisasContext *ctx, arg_SCMPU *a) |
| { |
| gen_helper_scmpu(cpu_env); |
| return true; |
| } |
| |
| /* smovu */ |
| static bool trans_SMOVU(DisasContext *ctx, arg_SMOVU *a) |
| { |
| gen_helper_smovu(cpu_env); |
| return true; |
| } |
| |
| /* smovf */ |
| static bool trans_SMOVF(DisasContext *ctx, arg_SMOVF *a) |
| { |
| gen_helper_smovf(cpu_env); |
| return true; |
| } |
| |
| /* smovb */ |
| static bool trans_SMOVB(DisasContext *ctx, arg_SMOVB *a) |
| { |
| gen_helper_smovb(cpu_env); |
| return true; |
| } |
| |
| #define STRING(op) \ |
| do { \ |
| TCGv size = tcg_const_i32(a->sz); \ |
| gen_helper_##op(cpu_env, size); \ |
| tcg_temp_free(size); \ |
| } while (0) |
| |
| /* suntile.<bwl> */ |
| static bool trans_SUNTIL(DisasContext *ctx, arg_SUNTIL *a) |
| { |
| STRING(suntil); |
| return true; |
| } |
| |
| /* swhile.<bwl> */ |
| static bool trans_SWHILE(DisasContext *ctx, arg_SWHILE *a) |
| { |
| STRING(swhile); |
| return true; |
| } |
| /* sstr.<bwl> */ |
| static bool trans_SSTR(DisasContext *ctx, arg_SSTR *a) |
| { |
| STRING(sstr); |
| return true; |
| } |
| |
| /* rmpa.<bwl> */ |
| static bool trans_RMPA(DisasContext *ctx, arg_RMPA *a) |
| { |
| STRING(rmpa); |
| return true; |
| } |
| |
| static void rx_mul64hi(TCGv_i64 ret, int rs, int rs2) |
| { |
| TCGv_i64 tmp0, tmp1; |
| tmp0 = tcg_temp_new_i64(); |
| tmp1 = tcg_temp_new_i64(); |
| tcg_gen_ext_i32_i64(tmp0, cpu_regs[rs]); |
| tcg_gen_sari_i64(tmp0, tmp0, 16); |
| tcg_gen_ext_i32_i64(tmp1, cpu_regs[rs2]); |
| tcg_gen_sari_i64(tmp1, tmp1, 16); |
| tcg_gen_mul_i64(ret, tmp0, tmp1); |
| tcg_gen_shli_i64(ret, ret, 16); |
| tcg_temp_free_i64(tmp0); |
| tcg_temp_free_i64(tmp1); |
| } |
| |
| static void rx_mul64lo(TCGv_i64 ret, int rs, int rs2) |
| { |
| TCGv_i64 tmp0, tmp1; |
| tmp0 = tcg_temp_new_i64(); |
| tmp1 = tcg_temp_new_i64(); |
| tcg_gen_ext_i32_i64(tmp0, cpu_regs[rs]); |
| tcg_gen_ext16s_i64(tmp0, tmp0); |
| tcg_gen_ext_i32_i64(tmp1, cpu_regs[rs2]); |
| tcg_gen_ext16s_i64(tmp1, tmp1); |
| tcg_gen_mul_i64(ret, tmp0, tmp1); |
| tcg_gen_shli_i64(ret, ret, 16); |
| tcg_temp_free_i64(tmp0); |
| tcg_temp_free_i64(tmp1); |
| } |
| |
| /* mulhi rs,rs2 */ |
| static bool trans_MULHI(DisasContext *ctx, arg_MULHI *a) |
| { |
| rx_mul64hi(cpu_acc, a->rs, a->rs2); |
| return true; |
| } |
| |
| /* mullo rs,rs2 */ |
| static bool trans_MULLO(DisasContext *ctx, arg_MULLO *a) |
| { |
| rx_mul64lo(cpu_acc, a->rs, a->rs2); |
| return true; |
| } |
| |
| /* machi rs,rs2 */ |
| static bool trans_MACHI(DisasContext *ctx, arg_MACHI *a) |
| { |
| TCGv_i64 tmp; |
| tmp = tcg_temp_new_i64(); |
| rx_mul64hi(tmp, a->rs, a->rs2); |
| tcg_gen_add_i64(cpu_acc, cpu_acc, tmp); |
| tcg_temp_free_i64(tmp); |
| return true; |
| } |
| |
| /* maclo rs,rs2 */ |
| static bool trans_MACLO(DisasContext *ctx, arg_MACLO *a) |
| { |
| TCGv_i64 tmp; |
| tmp = tcg_temp_new_i64(); |
| rx_mul64lo(tmp, a->rs, a->rs2); |
| tcg_gen_add_i64(cpu_acc, cpu_acc, tmp); |
| tcg_temp_free_i64(tmp); |
| return true; |
| } |
| |
| /* mvfachi rd */ |
| static bool trans_MVFACHI(DisasContext *ctx, arg_MVFACHI *a) |
| { |
| tcg_gen_extrh_i64_i32(cpu_regs[a->rd], cpu_acc); |
| return true; |
| } |
| |
| /* mvfacmi rd */ |
| static bool trans_MVFACMI(DisasContext *ctx, arg_MVFACMI *a) |
| { |
| TCGv_i64 rd64; |
| rd64 = tcg_temp_new_i64(); |
| tcg_gen_extract_i64(rd64, cpu_acc, 16, 32); |
| tcg_gen_extrl_i64_i32(cpu_regs[a->rd], rd64); |
| tcg_temp_free_i64(rd64); |
| return true; |
| } |
| |
| /* mvtachi rs */ |
| static bool trans_MVTACHI(DisasContext *ctx, arg_MVTACHI *a) |
| { |
| TCGv_i64 rs64; |
| rs64 = tcg_temp_new_i64(); |
| tcg_gen_extu_i32_i64(rs64, cpu_regs[a->rs]); |
| tcg_gen_deposit_i64(cpu_acc, cpu_acc, rs64, 32, 32); |
| tcg_temp_free_i64(rs64); |
| return true; |
| } |
| |
| /* mvtaclo rs */ |
| static bool trans_MVTACLO(DisasContext *ctx, arg_MVTACLO *a) |
| { |
| TCGv_i64 rs64; |
| rs64 = tcg_temp_new_i64(); |
| tcg_gen_extu_i32_i64(rs64, cpu_regs[a->rs]); |
| tcg_gen_deposit_i64(cpu_acc, cpu_acc, rs64, 0, 32); |
| tcg_temp_free_i64(rs64); |
| return true; |
| } |
| |
| /* racw #imm */ |
| static bool trans_RACW(DisasContext *ctx, arg_RACW *a) |
| { |
| TCGv imm = tcg_const_i32(a->imm + 1); |
| gen_helper_racw(cpu_env, imm); |
| tcg_temp_free(imm); |
| return true; |
| } |
| |
| /* sat rd */ |
| static bool trans_SAT(DisasContext *ctx, arg_SAT *a) |
| { |
| TCGv tmp, z; |
| tmp = tcg_temp_new(); |
| z = tcg_const_i32(0); |
| /* S == 1 -> 0xffffffff / S == 0 -> 0x00000000 */ |
| tcg_gen_sari_i32(tmp, cpu_psw_s, 31); |
| /* S == 1 -> 0x7fffffff / S == 0 -> 0x80000000 */ |
| tcg_gen_xori_i32(tmp, tmp, 0x80000000); |
| tcg_gen_movcond_i32(TCG_COND_LT, cpu_regs[a->rd], |
| cpu_psw_o, z, tmp, cpu_regs[a->rd]); |
| tcg_temp_free(tmp); |
| tcg_temp_free(z); |
| return true; |
| } |
| |
| /* satr */ |
| static bool trans_SATR(DisasContext *ctx, arg_SATR *a) |
| { |
| gen_helper_satr(cpu_env); |
| return true; |
| } |
| |
| #define cat3(a, b, c) a##b##c |
| #define FOP(name, op) \ |
| static bool cat3(trans_, name, _ir)(DisasContext *ctx, \ |
| cat3(arg_, name, _ir) * a) \ |
| { \ |
| TCGv imm = tcg_const_i32(li(ctx, 0)); \ |
| gen_helper_##op(cpu_regs[a->rd], cpu_env, \ |
| cpu_regs[a->rd], imm); \ |
| tcg_temp_free(imm); \ |
| return true; \ |
| } \ |
| static bool cat3(trans_, name, _mr)(DisasContext *ctx, \ |
| cat3(arg_, name, _mr) * a) \ |
| { \ |
| TCGv val, mem; \ |
| mem = tcg_temp_new(); \ |
| val = rx_load_source(ctx, mem, a->ld, MO_32, a->rs); \ |
| gen_helper_##op(cpu_regs[a->rd], cpu_env, \ |
| cpu_regs[a->rd], val); \ |
| tcg_temp_free(mem); \ |
| return true; \ |
| } |
| |
| #define FCONVOP(name, op) \ |
| static bool trans_##name(DisasContext *ctx, arg_##name * a) \ |
| { \ |
| TCGv val, mem; \ |
| mem = tcg_temp_new(); \ |
| val = rx_load_source(ctx, mem, a->ld, MO_32, a->rs); \ |
| gen_helper_##op(cpu_regs[a->rd], cpu_env, val); \ |
| tcg_temp_free(mem); \ |
| return true; \ |
| } |
| |
| FOP(FADD, fadd) |
| FOP(FSUB, fsub) |
| FOP(FMUL, fmul) |
| FOP(FDIV, fdiv) |
| |
| /* fcmp #imm, rd */ |
| static bool trans_FCMP_ir(DisasContext *ctx, arg_FCMP_ir * a) |
| { |
| TCGv imm = tcg_const_i32(li(ctx, 0)); |
| gen_helper_fcmp(cpu_env, cpu_regs[a->rd], imm); |
| tcg_temp_free(imm); |
| return true; |
| } |
| |
| /* fcmp dsp[rs], rd */ |
| /* fcmp rs, rd */ |
| static bool trans_FCMP_mr(DisasContext *ctx, arg_FCMP_mr *a) |
| { |
| TCGv val, mem; |
| mem = tcg_temp_new(); |
| val = rx_load_source(ctx, mem, a->ld, MO_32, a->rs); |
| gen_helper_fcmp(cpu_env, cpu_regs[a->rd], val); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| FCONVOP(FTOI, ftoi) |
| FCONVOP(ROUND, round) |
| |
| /* itof rs, rd */ |
| /* itof dsp[rs], rd */ |
| static bool trans_ITOF(DisasContext *ctx, arg_ITOF * a) |
| { |
| TCGv val, mem; |
| mem = tcg_temp_new(); |
| val = rx_load_source(ctx, mem, a->ld, a->mi, a->rs); |
| gen_helper_itof(cpu_regs[a->rd], cpu_env, val); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| static void rx_bsetm(TCGv mem, TCGv mask) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| rx_gen_ld(MO_8, val, mem); |
| tcg_gen_or_i32(val, val, mask); |
| rx_gen_st(MO_8, val, mem); |
| tcg_temp_free(val); |
| } |
| |
| static void rx_bclrm(TCGv mem, TCGv mask) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| rx_gen_ld(MO_8, val, mem); |
| tcg_gen_andc_i32(val, val, mask); |
| rx_gen_st(MO_8, val, mem); |
| tcg_temp_free(val); |
| } |
| |
| static void rx_btstm(TCGv mem, TCGv mask) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| rx_gen_ld(MO_8, val, mem); |
| tcg_gen_and_i32(val, val, mask); |
| tcg_gen_setcondi_i32(TCG_COND_NE, cpu_psw_c, val, 0); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_c); |
| tcg_temp_free(val); |
| } |
| |
| static void rx_bnotm(TCGv mem, TCGv mask) |
| { |
| TCGv val; |
| val = tcg_temp_new(); |
| rx_gen_ld(MO_8, val, mem); |
| tcg_gen_xor_i32(val, val, mask); |
| rx_gen_st(MO_8, val, mem); |
| tcg_temp_free(val); |
| } |
| |
| static void rx_bsetr(TCGv reg, TCGv mask) |
| { |
| tcg_gen_or_i32(reg, reg, mask); |
| } |
| |
| static void rx_bclrr(TCGv reg, TCGv mask) |
| { |
| tcg_gen_andc_i32(reg, reg, mask); |
| } |
| |
| static inline void rx_btstr(TCGv reg, TCGv mask) |
| { |
| TCGv t0; |
| t0 = tcg_temp_new(); |
| tcg_gen_and_i32(t0, reg, mask); |
| tcg_gen_setcondi_i32(TCG_COND_NE, cpu_psw_c, t0, 0); |
| tcg_gen_mov_i32(cpu_psw_z, cpu_psw_c); |
| tcg_temp_free(t0); |
| } |
| |
| static inline void rx_bnotr(TCGv reg, TCGv mask) |
| { |
| tcg_gen_xor_i32(reg, reg, mask); |
| } |
| |
| #define BITOP(name, op) \ |
| static bool cat3(trans_, name, _im)(DisasContext *ctx, \ |
| cat3(arg_, name, _im) * a) \ |
| { \ |
| TCGv mask, mem, addr; \ |
| mem = tcg_temp_new(); \ |
| mask = tcg_const_i32(1 << a->imm); \ |
| addr = rx_index_addr(ctx, mem, a->ld, MO_8, a->rs); \ |
| cat3(rx_, op, m)(addr, mask); \ |
| tcg_temp_free(mask); \ |
| tcg_temp_free(mem); \ |
| return true; \ |
| } \ |
| static bool cat3(trans_, name, _ir)(DisasContext *ctx, \ |
| cat3(arg_, name, _ir) * a) \ |
| { \ |
| TCGv mask; \ |
| mask = tcg_const_i32(1 << a->imm); \ |
| cat3(rx_, op, r)(cpu_regs[a->rd], mask); \ |
| tcg_temp_free(mask); \ |
| return true; \ |
| } \ |
| static bool cat3(trans_, name, _rr)(DisasContext *ctx, \ |
| cat3(arg_, name, _rr) * a) \ |
| { \ |
| TCGv mask, b; \ |
| mask = tcg_const_i32(1); \ |
| b = tcg_temp_new(); \ |
| tcg_gen_andi_i32(b, cpu_regs[a->rs], 31); \ |
| tcg_gen_shl_i32(mask, mask, b); \ |
| cat3(rx_, op, r)(cpu_regs[a->rd], mask); \ |
| tcg_temp_free(mask); \ |
| tcg_temp_free(b); \ |
| return true; \ |
| } \ |
| static bool cat3(trans_, name, _rm)(DisasContext *ctx, \ |
| cat3(arg_, name, _rm) * a) \ |
| { \ |
| TCGv mask, mem, addr, b; \ |
| mask = tcg_const_i32(1); \ |
| b = tcg_temp_new(); \ |
| tcg_gen_andi_i32(b, cpu_regs[a->rd], 7); \ |
| tcg_gen_shl_i32(mask, mask, b); \ |
| mem = tcg_temp_new(); \ |
| addr = rx_index_addr(ctx, mem, a->ld, MO_8, a->rs); \ |
| cat3(rx_, op, m)(addr, mask); \ |
| tcg_temp_free(mem); \ |
| tcg_temp_free(mask); \ |
| tcg_temp_free(b); \ |
| return true; \ |
| } |
| |
| BITOP(BSET, bset) |
| BITOP(BCLR, bclr) |
| BITOP(BTST, btst) |
| BITOP(BNOT, bnot) |
| |
| static inline void bmcnd_op(TCGv val, TCGCond cond, int pos) |
| { |
| TCGv bit; |
| DisasCompare dc; |
| dc.temp = tcg_temp_new(); |
| bit = tcg_temp_new(); |
| psw_cond(&dc, cond); |
| tcg_gen_andi_i32(val, val, ~(1 << pos)); |
| tcg_gen_setcondi_i32(dc.cond, bit, dc.value, 0); |
| tcg_gen_deposit_i32(val, val, bit, pos, 1); |
| tcg_temp_free(bit); |
| tcg_temp_free(dc.temp); |
| } |
| |
| /* bmcnd #imm, dsp[rd] */ |
| static bool trans_BMCnd_im(DisasContext *ctx, arg_BMCnd_im *a) |
| { |
| TCGv val, mem, addr; |
| val = tcg_temp_new(); |
| mem = tcg_temp_new(); |
| addr = rx_index_addr(ctx, mem, a->ld, MO_8, a->rd); |
| rx_gen_ld(MO_8, val, addr); |
| bmcnd_op(val, a->cd, a->imm); |
| rx_gen_st(MO_8, val, addr); |
| tcg_temp_free(val); |
| tcg_temp_free(mem); |
| return true; |
| } |
| |
| /* bmcond #imm, rd */ |
| static bool trans_BMCnd_ir(DisasContext *ctx, arg_BMCnd_ir *a) |
| { |
| bmcnd_op(cpu_regs[a->rd], a->cd, a->imm); |
| return true; |
| } |
| |
| enum { |
| PSW_C = 0, |
| PSW_Z = 1, |
| PSW_S = 2, |
| PSW_O = 3, |
| PSW_I = 8, |
| PSW_U = 9, |
| }; |
| |
| static inline void clrsetpsw(DisasContext *ctx, int cb, int val) |
| { |
| if (cb < 8) { |
| switch (cb) { |
| case PSW_C: |
| tcg_gen_movi_i32(cpu_psw_c, val); |
| break; |
| case PSW_Z: |
| tcg_gen_movi_i32(cpu_psw_z, val == 0); |
| break; |
| case PSW_S: |
| tcg_gen_movi_i32(cpu_psw_s, val ? -1 : 0); |
| break; |
| case PSW_O: |
| tcg_gen_movi_i32(cpu_psw_o, val << 31); |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, "Invalid distination %d", cb); |
| break; |
| } |
| } else if (is_privileged(ctx, 0)) { |
| switch (cb) { |
| case PSW_I: |
| tcg_gen_movi_i32(cpu_psw_i, val); |
| ctx->base.is_jmp = DISAS_UPDATE; |
| break; |
| case PSW_U: |
| tcg_gen_movi_i32(cpu_psw_u, val); |
| break; |
| default: |
| qemu_log_mask(LOG_GUEST_ERROR, "Invalid distination %d", cb); |
| break; |
| } |
| } |
| } |
| |
| /* clrpsw psw */ |
| static bool trans_CLRPSW(DisasContext *ctx, arg_CLRPSW *a) |
| { |
| clrsetpsw(ctx, a->cb, 0); |
| return true; |
| } |
| |
| /* setpsw psw */ |
| static bool trans_SETPSW(DisasContext *ctx, arg_SETPSW *a) |
| { |
| clrsetpsw(ctx, a->cb, 1); |
| return true; |
| } |
| |
| /* mvtipl #imm */ |
| static bool trans_MVTIPL(DisasContext *ctx, arg_MVTIPL *a) |
| { |
| if (is_privileged(ctx, 1)) { |
| tcg_gen_movi_i32(cpu_psw_ipl, a->imm); |
| ctx->base.is_jmp = DISAS_UPDATE; |
| } |
| return true; |
| } |
| |
| /* mvtc #imm, rd */ |
| static bool trans_MVTC_i(DisasContext *ctx, arg_MVTC_i *a) |
| { |
| TCGv imm; |
| |
| imm = tcg_const_i32(a->imm); |
| move_to_cr(ctx, imm, a->cr); |
| if (a->cr == 0 && is_privileged(ctx, 0)) { |
| ctx->base.is_jmp = DISAS_UPDATE; |
| } |
| tcg_temp_free(imm); |
| return true; |
| } |
| |
| /* mvtc rs, rd */ |
| static bool trans_MVTC_r(DisasContext *ctx, arg_MVTC_r *a) |
| { |
| move_to_cr(ctx, cpu_regs[a->rs], a->cr); |
| if (a->cr == 0 && is_privileged(ctx, 0)) { |
| ctx->base.is_jmp = DISAS_UPDATE; |
| } |
| return true; |
| } |
| |
| /* mvfc rs, rd */ |
| static bool trans_MVFC(DisasContext *ctx, arg_MVFC *a) |
| { |
| move_from_cr(cpu_regs[a->rd], a->cr, ctx->pc); |
| return true; |
| } |
| |
| /* rtfi */ |
| static bool trans_RTFI(DisasContext *ctx, arg_RTFI *a) |
| { |
| TCGv psw; |
| if (is_privileged(ctx, 1)) { |
| psw = tcg_temp_new(); |
| tcg_gen_mov_i32(cpu_pc, cpu_bpc); |
| tcg_gen_mov_i32(psw, cpu_bpsw); |
| gen_helper_set_psw_rte(cpu_env, psw); |
| ctx->base.is_jmp = DISAS_EXIT; |
| tcg_temp_free(psw); |
| } |
| return true; |
| } |
| |
| /* rte */ |
| static bool trans_RTE(DisasContext *ctx, arg_RTE *a) |
| { |
| TCGv psw; |
| if (is_privileged(ctx, 1)) { |
| psw = tcg_temp_new(); |
| pop(cpu_pc); |
| pop(psw); |
| gen_helper_set_psw_rte(cpu_env, psw); |
| ctx->base.is_jmp = DISAS_EXIT; |
| tcg_temp_free(psw); |
| } |
| return true; |
| } |
| |
| /* brk */ |
| static bool trans_BRK(DisasContext *ctx, arg_BRK *a) |
| { |
| tcg_gen_movi_i32(cpu_pc, ctx->base.pc_next); |
| gen_helper_rxbrk(cpu_env); |
| ctx->base.is_jmp = DISAS_NORETURN; |
| return true; |
| } |
| |
| /* int #imm */ |
| static bool trans_INT(DisasContext *ctx, arg_INT *a) |
| { |
| TCGv vec; |
| |
| tcg_debug_assert(a->imm < 0x100); |
| vec = tcg_const_i32(a->imm); |
| tcg_gen_movi_i32(cpu_pc, ctx->base.pc_next); |
| gen_helper_rxint(cpu_env, vec); |
| tcg_temp_free(vec); |
| ctx->base.is_jmp = DISAS_NORETURN; |
| return true; |
| } |
| |
| /* wait */ |
| static bool trans_WAIT(DisasContext *ctx, arg_WAIT *a) |
| { |
| if (is_privileged(ctx, 1)) { |
| tcg_gen_addi_i32(cpu_pc, cpu_pc, 2); |
| gen_helper_wait(cpu_env); |
| } |
| return true; |
| } |
| |
| static void rx_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs) |
| { |
| CPURXState *env = cs->env_ptr; |
| DisasContext *ctx = container_of(dcbase, DisasContext, base); |
| ctx->env = env; |
| } |
| |
| static void rx_tr_tb_start(DisasContextBase *dcbase, CPUState *cs) |
| { |
| } |
| |
| static void rx_tr_insn_start(DisasContextBase *dcbase, CPUState *cs) |
| { |
| DisasContext *ctx = container_of(dcbase, DisasContext, base); |
| |
| tcg_gen_insn_start(ctx->base.pc_next); |
| } |
| |
| static void rx_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs) |
| { |
| DisasContext *ctx = container_of(dcbase, DisasContext, base); |
| uint32_t insn; |
| |
| ctx->pc = ctx->base.pc_next; |
| insn = decode_load(ctx); |
| if (!decode(ctx, insn)) { |
| gen_helper_raise_illegal_instruction(cpu_env); |
| } |
| } |
| |
| static void rx_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs) |
| { |
| DisasContext *ctx = container_of(dcbase, DisasContext, base); |
| |
| switch (ctx->base.is_jmp) { |
| case DISAS_NEXT: |
| case DISAS_TOO_MANY: |
| gen_goto_tb(ctx, 0, dcbase->pc_next); |
| break; |
| case DISAS_JUMP: |
| if (ctx->base.singlestep_enabled) { |
| gen_helper_debug(cpu_env); |
| } else { |
| tcg_gen_lookup_and_goto_ptr(); |
| } |
| break; |
| case DISAS_UPDATE: |
| tcg_gen_movi_i32(cpu_pc, ctx->base.pc_next); |
| /* fall through */ |
| case DISAS_EXIT: |
| tcg_gen_exit_tb(NULL, 0); |
| break; |
| case DISAS_NORETURN: |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| |
| static void rx_tr_disas_log(const DisasContextBase *dcbase, CPUState *cs) |
| { |
| qemu_log("IN:\n"); /* , lookup_symbol(dcbase->pc_first)); */ |
| log_target_disas(cs, dcbase->pc_first, dcbase->tb->size); |
| } |
| |
| static const TranslatorOps rx_tr_ops = { |
| .init_disas_context = rx_tr_init_disas_context, |
| .tb_start = rx_tr_tb_start, |
| .insn_start = rx_tr_insn_start, |
| .translate_insn = rx_tr_translate_insn, |
| .tb_stop = rx_tr_tb_stop, |
| .disas_log = rx_tr_disas_log, |
| }; |
| |
| void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns) |
| { |
| DisasContext dc; |
| |
| translator_loop(&rx_tr_ops, &dc.base, cs, tb, max_insns); |
| } |
| |
| void restore_state_to_opc(CPURXState *env, TranslationBlock *tb, |
| target_ulong *data) |
| { |
| env->pc = data[0]; |
| } |
| |
| #define ALLOC_REGISTER(sym, name) \ |
| cpu_##sym = tcg_global_mem_new_i32(cpu_env, \ |
| offsetof(CPURXState, sym), name) |
| |
| void rx_translate_init(void) |
| { |
| static const char * const regnames[NUM_REGS] = { |
| "R0", "R1", "R2", "R3", "R4", "R5", "R6", "R7", |
| "R8", "R9", "R10", "R11", "R12", "R13", "R14", "R15" |
| }; |
| int i; |
| |
| for (i = 0; i < NUM_REGS; i++) { |
| cpu_regs[i] = tcg_global_mem_new_i32(cpu_env, |
| offsetof(CPURXState, regs[i]), |
| regnames[i]); |
| } |
| ALLOC_REGISTER(pc, "PC"); |
| ALLOC_REGISTER(psw_o, "PSW(O)"); |
| ALLOC_REGISTER(psw_s, "PSW(S)"); |
| ALLOC_REGISTER(psw_z, "PSW(Z)"); |
| ALLOC_REGISTER(psw_c, "PSW(C)"); |
| ALLOC_REGISTER(psw_u, "PSW(U)"); |
| ALLOC_REGISTER(psw_i, "PSW(I)"); |
| ALLOC_REGISTER(psw_pm, "PSW(PM)"); |
| ALLOC_REGISTER(psw_ipl, "PSW(IPL)"); |
| ALLOC_REGISTER(usp, "USP"); |
| ALLOC_REGISTER(fpsw, "FPSW"); |
| ALLOC_REGISTER(bpsw, "BPSW"); |
| ALLOC_REGISTER(bpc, "BPC"); |
| ALLOC_REGISTER(isp, "ISP"); |
| ALLOC_REGISTER(fintv, "FINTV"); |
| ALLOC_REGISTER(intb, "INTB"); |
| cpu_acc = tcg_global_mem_new_i64(cpu_env, |
| offsetof(CPURXState, acc), "ACC"); |
| } |