| /* |
| * QEMU TCG support -- s390x vector instruction translation functions |
| * |
| * Copyright (C) 2019 Red Hat Inc |
| * |
| * Authors: |
| * David Hildenbrand <david@redhat.com> |
| * |
| * This work is licensed under the terms of the GNU GPL, version 2 or later. |
| * See the COPYING file in the top-level directory. |
| */ |
| |
| /* |
| * For most instructions that use the same element size for reads and |
| * writes, we can use real gvec vector expansion, which potantially uses |
| * real host vector instructions. As they only work up to 64 bit elements, |
| * 128 bit elements (vector is a single element) have to be handled |
| * differently. Operations that are too complicated to encode via TCG ops |
| * are handled via gvec ool (out-of-line) handlers. |
| * |
| * As soon as instructions use different element sizes for reads and writes |
| * or access elements "out of their element scope" we expand them manually |
| * in fancy loops, as gvec expansion does not deal with actual element |
| * numbers and does also not support access to other elements. |
| * |
| * 128 bit elements: |
| * As we only have i32/i64, such elements have to be loaded into two |
| * i64 values and can then be processed e.g. by tcg_gen_add2_i64. |
| * |
| * Sizes: |
| * On s390x, the operand size (oprsz) and the maximum size (maxsz) are |
| * always 16 (128 bit). What gvec code calls "vece", s390x calls "es", |
| * a.k.a. "element size". These values nicely map to MO_8 ... MO_64. Only |
| * 128 bit element size has to be treated in a special way (MO_64 + 1). |
| * We will use ES_* instead of MO_* for this reason in this file. |
| * |
| * CC handling: |
| * As gvec ool-helpers can currently not return values (besides via |
| * pointers like vectors or cpu_env), whenever we have to set the CC and |
| * can't conclude the value from the result vector, we will directly |
| * set it in "env->cc_op" and mark it as static via set_cc_static()". |
| * Whenever this is done, the helper writes globals (cc_op). |
| */ |
| |
| #define NUM_VEC_ELEMENT_BYTES(es) (1 << (es)) |
| #define NUM_VEC_ELEMENTS(es) (16 / NUM_VEC_ELEMENT_BYTES(es)) |
| #define NUM_VEC_ELEMENT_BITS(es) (NUM_VEC_ELEMENT_BYTES(es) * BITS_PER_BYTE) |
| |
| #define ES_8 MO_8 |
| #define ES_16 MO_16 |
| #define ES_32 MO_32 |
| #define ES_64 MO_64 |
| #define ES_128 4 |
| |
| /* Floating-Point Format */ |
| #define FPF_SHORT 2 |
| #define FPF_LONG 3 |
| #define FPF_EXT 4 |
| |
| static inline bool valid_vec_element(uint8_t enr, MemOp es) |
| { |
| return !(enr & ~(NUM_VEC_ELEMENTS(es) - 1)); |
| } |
| |
| static void read_vec_element_i64(TCGv_i64 dst, uint8_t reg, uint8_t enr, |
| MemOp memop) |
| { |
| const int offs = vec_reg_offset(reg, enr, memop & MO_SIZE); |
| |
| switch ((unsigned)memop) { |
| case ES_8: |
| tcg_gen_ld8u_i64(dst, cpu_env, offs); |
| break; |
| case ES_16: |
| tcg_gen_ld16u_i64(dst, cpu_env, offs); |
| break; |
| case ES_32: |
| tcg_gen_ld32u_i64(dst, cpu_env, offs); |
| break; |
| case ES_8 | MO_SIGN: |
| tcg_gen_ld8s_i64(dst, cpu_env, offs); |
| break; |
| case ES_16 | MO_SIGN: |
| tcg_gen_ld16s_i64(dst, cpu_env, offs); |
| break; |
| case ES_32 | MO_SIGN: |
| tcg_gen_ld32s_i64(dst, cpu_env, offs); |
| break; |
| case ES_64: |
| case ES_64 | MO_SIGN: |
| tcg_gen_ld_i64(dst, cpu_env, offs); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| |
| static void read_vec_element_i32(TCGv_i32 dst, uint8_t reg, uint8_t enr, |
| MemOp memop) |
| { |
| const int offs = vec_reg_offset(reg, enr, memop & MO_SIZE); |
| |
| switch (memop) { |
| case ES_8: |
| tcg_gen_ld8u_i32(dst, cpu_env, offs); |
| break; |
| case ES_16: |
| tcg_gen_ld16u_i32(dst, cpu_env, offs); |
| break; |
| case ES_8 | MO_SIGN: |
| tcg_gen_ld8s_i32(dst, cpu_env, offs); |
| break; |
| case ES_16 | MO_SIGN: |
| tcg_gen_ld16s_i32(dst, cpu_env, offs); |
| break; |
| case ES_32: |
| case ES_32 | MO_SIGN: |
| tcg_gen_ld_i32(dst, cpu_env, offs); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| |
| static void write_vec_element_i64(TCGv_i64 src, int reg, uint8_t enr, |
| MemOp memop) |
| { |
| const int offs = vec_reg_offset(reg, enr, memop & MO_SIZE); |
| |
| switch (memop) { |
| case ES_8: |
| tcg_gen_st8_i64(src, cpu_env, offs); |
| break; |
| case ES_16: |
| tcg_gen_st16_i64(src, cpu_env, offs); |
| break; |
| case ES_32: |
| tcg_gen_st32_i64(src, cpu_env, offs); |
| break; |
| case ES_64: |
| tcg_gen_st_i64(src, cpu_env, offs); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| |
| static void write_vec_element_i32(TCGv_i32 src, int reg, uint8_t enr, |
| MemOp memop) |
| { |
| const int offs = vec_reg_offset(reg, enr, memop & MO_SIZE); |
| |
| switch (memop) { |
| case ES_8: |
| tcg_gen_st8_i32(src, cpu_env, offs); |
| break; |
| case ES_16: |
| tcg_gen_st16_i32(src, cpu_env, offs); |
| break; |
| case ES_32: |
| tcg_gen_st_i32(src, cpu_env, offs); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| |
| static void get_vec_element_ptr_i64(TCGv_ptr ptr, uint8_t reg, TCGv_i64 enr, |
| uint8_t es) |
| { |
| TCGv_i64 tmp = tcg_temp_new_i64(); |
| |
| /* mask off invalid parts from the element nr */ |
| tcg_gen_andi_i64(tmp, enr, NUM_VEC_ELEMENTS(es) - 1); |
| |
| /* convert it to an element offset relative to cpu_env (vec_reg_offset() */ |
| tcg_gen_shli_i64(tmp, tmp, es); |
| #if !HOST_BIG_ENDIAN |
| tcg_gen_xori_i64(tmp, tmp, 8 - NUM_VEC_ELEMENT_BYTES(es)); |
| #endif |
| tcg_gen_addi_i64(tmp, tmp, vec_full_reg_offset(reg)); |
| |
| /* generate the final ptr by adding cpu_env */ |
| tcg_gen_trunc_i64_ptr(ptr, tmp); |
| tcg_gen_add_ptr(ptr, ptr, cpu_env); |
| } |
| |
| #define gen_gvec_2(v1, v2, gen) \ |
| tcg_gen_gvec_2(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| 16, 16, gen) |
| #define gen_gvec_2s(v1, v2, c, gen) \ |
| tcg_gen_gvec_2s(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| 16, 16, c, gen) |
| #define gen_gvec_2_ool(v1, v2, data, fn) \ |
| tcg_gen_gvec_2_ool(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| 16, 16, data, fn) |
| #define gen_gvec_2i_ool(v1, v2, c, data, fn) \ |
| tcg_gen_gvec_2i_ool(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| c, 16, 16, data, fn) |
| #define gen_gvec_2_ptr(v1, v2, ptr, data, fn) \ |
| tcg_gen_gvec_2_ptr(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| ptr, 16, 16, data, fn) |
| #define gen_gvec_3(v1, v2, v3, gen) \ |
| tcg_gen_gvec_3(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), 16, 16, gen) |
| #define gen_gvec_3_ool(v1, v2, v3, data, fn) \ |
| tcg_gen_gvec_3_ool(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), 16, 16, data, fn) |
| #define gen_gvec_3_ptr(v1, v2, v3, ptr, data, fn) \ |
| tcg_gen_gvec_3_ptr(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), ptr, 16, 16, data, fn) |
| #define gen_gvec_3i(v1, v2, v3, c, gen) \ |
| tcg_gen_gvec_3i(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), 16, 16, c, gen) |
| #define gen_gvec_4(v1, v2, v3, v4, gen) \ |
| tcg_gen_gvec_4(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), vec_full_reg_offset(v4), \ |
| 16, 16, gen) |
| #define gen_gvec_4_ool(v1, v2, v3, v4, data, fn) \ |
| tcg_gen_gvec_4_ool(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), vec_full_reg_offset(v4), \ |
| 16, 16, data, fn) |
| #define gen_gvec_4_ptr(v1, v2, v3, v4, ptr, data, fn) \ |
| tcg_gen_gvec_4_ptr(vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), vec_full_reg_offset(v4), \ |
| ptr, 16, 16, data, fn) |
| #define gen_gvec_dup_i64(es, v1, c) \ |
| tcg_gen_gvec_dup_i64(es, vec_full_reg_offset(v1), 16, 16, c) |
| #define gen_gvec_mov(v1, v2) \ |
| tcg_gen_gvec_mov(0, vec_full_reg_offset(v1), vec_full_reg_offset(v2), 16, \ |
| 16) |
| #define gen_gvec_dup_imm(es, v1, c) \ |
| tcg_gen_gvec_dup_imm(es, vec_full_reg_offset(v1), 16, 16, c); |
| #define gen_gvec_fn_2(fn, es, v1, v2) \ |
| tcg_gen_gvec_##fn(es, vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| 16, 16) |
| #define gen_gvec_fn_2i(fn, es, v1, v2, c) \ |
| tcg_gen_gvec_##fn(es, vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| c, 16, 16) |
| #define gen_gvec_fn_2s(fn, es, v1, v2, s) \ |
| tcg_gen_gvec_##fn(es, vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| s, 16, 16) |
| #define gen_gvec_fn_3(fn, es, v1, v2, v3) \ |
| tcg_gen_gvec_##fn(es, vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), 16, 16) |
| #define gen_gvec_fn_4(fn, es, v1, v2, v3, v4) \ |
| tcg_gen_gvec_##fn(es, vec_full_reg_offset(v1), vec_full_reg_offset(v2), \ |
| vec_full_reg_offset(v3), vec_full_reg_offset(v4), 16, 16) |
| |
| /* |
| * Helper to carry out a 128 bit vector computation using 2 i64 values per |
| * vector. |
| */ |
| typedef void (*gen_gvec128_3_i64_fn)(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, |
| TCGv_i64 ah, TCGv_i64 bl, TCGv_i64 bh); |
| static void gen_gvec128_3_i64(gen_gvec128_3_i64_fn fn, uint8_t d, uint8_t a, |
| uint8_t b) |
| { |
| TCGv_i64 dh = tcg_temp_new_i64(); |
| TCGv_i64 dl = tcg_temp_new_i64(); |
| TCGv_i64 ah = tcg_temp_new_i64(); |
| TCGv_i64 al = tcg_temp_new_i64(); |
| TCGv_i64 bh = tcg_temp_new_i64(); |
| TCGv_i64 bl = tcg_temp_new_i64(); |
| |
| read_vec_element_i64(ah, a, 0, ES_64); |
| read_vec_element_i64(al, a, 1, ES_64); |
| read_vec_element_i64(bh, b, 0, ES_64); |
| read_vec_element_i64(bl, b, 1, ES_64); |
| fn(dl, dh, al, ah, bl, bh); |
| write_vec_element_i64(dh, d, 0, ES_64); |
| write_vec_element_i64(dl, d, 1, ES_64); |
| } |
| |
| typedef void (*gen_gvec128_4_i64_fn)(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, |
| TCGv_i64 ah, TCGv_i64 bl, TCGv_i64 bh, |
| TCGv_i64 cl, TCGv_i64 ch); |
| static void gen_gvec128_4_i64(gen_gvec128_4_i64_fn fn, uint8_t d, uint8_t a, |
| uint8_t b, uint8_t c) |
| { |
| TCGv_i64 dh = tcg_temp_new_i64(); |
| TCGv_i64 dl = tcg_temp_new_i64(); |
| TCGv_i64 ah = tcg_temp_new_i64(); |
| TCGv_i64 al = tcg_temp_new_i64(); |
| TCGv_i64 bh = tcg_temp_new_i64(); |
| TCGv_i64 bl = tcg_temp_new_i64(); |
| TCGv_i64 ch = tcg_temp_new_i64(); |
| TCGv_i64 cl = tcg_temp_new_i64(); |
| |
| read_vec_element_i64(ah, a, 0, ES_64); |
| read_vec_element_i64(al, a, 1, ES_64); |
| read_vec_element_i64(bh, b, 0, ES_64); |
| read_vec_element_i64(bl, b, 1, ES_64); |
| read_vec_element_i64(ch, c, 0, ES_64); |
| read_vec_element_i64(cl, c, 1, ES_64); |
| fn(dl, dh, al, ah, bl, bh, cl, ch); |
| write_vec_element_i64(dh, d, 0, ES_64); |
| write_vec_element_i64(dl, d, 1, ES_64); |
| } |
| |
| static void gen_addi2_i64(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, TCGv_i64 ah, |
| uint64_t b) |
| { |
| TCGv_i64 bl = tcg_constant_i64(b); |
| TCGv_i64 bh = tcg_constant_i64(0); |
| |
| tcg_gen_add2_i64(dl, dh, al, ah, bl, bh); |
| } |
| |
| static DisasJumpType op_vbperm(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_3_ool(get_field(s, v1), get_field(s, v2), get_field(s, v3), 0, |
| gen_helper_gvec_vbperm); |
| |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vge(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = s->insn->data; |
| const uint8_t enr = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (!valid_vec_element(enr, es)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| read_vec_element_i64(tmp, get_field(s, v2), enr, es); |
| tcg_gen_add_i64(o->addr1, o->addr1, tmp); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 0); |
| |
| tcg_gen_qemu_ld_i64(tmp, o->addr1, get_mem_index(s), MO_TE | es); |
| write_vec_element_i64(tmp, get_field(s, v1), enr, es); |
| return DISAS_NEXT; |
| } |
| |
| static uint64_t generate_byte_mask(uint8_t mask) |
| { |
| uint64_t r = 0; |
| int i; |
| |
| for (i = 0; i < 8; i++) { |
| if ((mask >> i) & 1) { |
| r |= 0xffull << (i * 8); |
| } |
| } |
| return r; |
| } |
| |
| static DisasJumpType op_vgbm(DisasContext *s, DisasOps *o) |
| { |
| const uint16_t i2 = get_field(s, i2); |
| |
| if (i2 == (i2 & 0xff) * 0x0101) { |
| /* |
| * Masks for both 64 bit elements of the vector are the same. |
| * Trust tcg to produce a good constant loading. |
| */ |
| gen_gvec_dup_imm(ES_64, get_field(s, v1), |
| generate_byte_mask(i2 & 0xff)); |
| } else { |
| TCGv_i64 t = tcg_temp_new_i64(); |
| |
| tcg_gen_movi_i64(t, generate_byte_mask(i2 >> 8)); |
| write_vec_element_i64(t, get_field(s, v1), 0, ES_64); |
| tcg_gen_movi_i64(t, generate_byte_mask(i2)); |
| write_vec_element_i64(t, get_field(s, v1), 1, ES_64); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vgm(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| const uint8_t bits = NUM_VEC_ELEMENT_BITS(es); |
| const uint8_t i2 = get_field(s, i2) & (bits - 1); |
| const uint8_t i3 = get_field(s, i3) & (bits - 1); |
| uint64_t mask = 0; |
| int i; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| /* generate the mask - take care of wrapping */ |
| for (i = i2; ; i = (i + 1) % bits) { |
| mask |= 1ull << (bits - i - 1); |
| if (i == i3) { |
| break; |
| } |
| } |
| |
| gen_gvec_dup_imm(es, get_field(s, v1), mask); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vl(DisasContext *s, DisasOps *o) |
| { |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| |
| tcg_gen_qemu_ld_i64(t0, o->addr1, get_mem_index(s), MO_TEUQ); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| tcg_gen_qemu_ld_i64(t1, o->addr1, get_mem_index(s), MO_TEUQ); |
| write_vec_element_i64(t0, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(t1, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlr(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_mov(get_field(s, v1), get_field(s, v2)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlrep(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(tmp, o->addr1, get_mem_index(s), MO_TE | es); |
| gen_gvec_dup_i64(es, get_field(s, v1), tmp); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlebr(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = s->insn->data; |
| const uint8_t enr = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (!valid_vec_element(enr, es)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(tmp, o->addr1, get_mem_index(s), MO_LE | es); |
| write_vec_element_i64(tmp, get_field(s, v1), enr, es); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlbrrep(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (es < ES_16 || es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(tmp, o->addr1, get_mem_index(s), MO_LE | es); |
| gen_gvec_dup_i64(es, get_field(s, v1), tmp); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vllebrz(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t m3 = get_field(s, m3); |
| TCGv_i64 tmp; |
| int es, lshift; |
| |
| switch (m3) { |
| case ES_16: |
| case ES_32: |
| case ES_64: |
| es = m3; |
| lshift = 0; |
| break; |
| case 6: |
| es = ES_32; |
| lshift = 32; |
| break; |
| default: |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(tmp, o->addr1, get_mem_index(s), MO_LE | es); |
| tcg_gen_shli_i64(tmp, tmp, lshift); |
| |
| write_vec_element_i64(tmp, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(tcg_constant_i64(0), get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlbr(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| TCGv_i64 t0, t1; |
| |
| if (es < ES_16 || es > ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| t0 = tcg_temp_new_i64(); |
| t1 = tcg_temp_new_i64(); |
| |
| |
| if (es == ES_128) { |
| tcg_gen_qemu_ld_i64(t1, o->addr1, get_mem_index(s), MO_LEUQ); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| tcg_gen_qemu_ld_i64(t0, o->addr1, get_mem_index(s), MO_LEUQ); |
| goto write; |
| } |
| |
| /* Begin with byte reversed doublewords... */ |
| tcg_gen_qemu_ld_i64(t0, o->addr1, get_mem_index(s), MO_LEUQ); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| tcg_gen_qemu_ld_i64(t1, o->addr1, get_mem_index(s), MO_LEUQ); |
| |
| /* |
| * For 16 and 32-bit elements, the doubleword bswap also reversed |
| * the order of the elements. Perform a larger order swap to put |
| * them back into place. For the 128-bit "element", finish the |
| * bswap by swapping the doublewords. |
| */ |
| switch (es) { |
| case ES_16: |
| tcg_gen_hswap_i64(t0, t0); |
| tcg_gen_hswap_i64(t1, t1); |
| break; |
| case ES_32: |
| tcg_gen_wswap_i64(t0, t0); |
| tcg_gen_wswap_i64(t1, t1); |
| break; |
| case ES_64: |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| write: |
| write_vec_element_i64(t0, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(t1, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vle(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = s->insn->data; |
| const uint8_t enr = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (!valid_vec_element(enr, es)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(tmp, o->addr1, get_mem_index(s), MO_TE | es); |
| write_vec_element_i64(tmp, get_field(s, v1), enr, es); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlei(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = s->insn->data; |
| const uint8_t enr = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (!valid_vec_element(enr, es)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_constant_i64((int16_t)get_field(s, i2)); |
| write_vec_element_i64(tmp, get_field(s, v1), enr, es); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vler(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| |
| if (es < ES_16 || es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| |
| /* Begin with the two doublewords swapped... */ |
| tcg_gen_qemu_ld_i64(t1, o->addr1, get_mem_index(s), MO_TEUQ); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| tcg_gen_qemu_ld_i64(t0, o->addr1, get_mem_index(s), MO_TEUQ); |
| |
| /* ... then swap smaller elements within the doublewords as required. */ |
| switch (es) { |
| case MO_16: |
| tcg_gen_hswap_i64(t1, t1); |
| tcg_gen_hswap_i64(t0, t0); |
| break; |
| case MO_32: |
| tcg_gen_wswap_i64(t1, t1); |
| tcg_gen_wswap_i64(t0, t0); |
| break; |
| case MO_64: |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| write_vec_element_i64(t0, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(t1, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlgv(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| TCGv_ptr ptr; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| /* fast path if we don't need the register content */ |
| if (!get_field(s, b2)) { |
| uint8_t enr = get_field(s, d2) & (NUM_VEC_ELEMENTS(es) - 1); |
| |
| read_vec_element_i64(o->out, get_field(s, v3), enr, es); |
| return DISAS_NEXT; |
| } |
| |
| ptr = tcg_temp_new_ptr(); |
| get_vec_element_ptr_i64(ptr, get_field(s, v3), o->addr1, es); |
| switch (es) { |
| case ES_8: |
| tcg_gen_ld8u_i64(o->out, ptr, 0); |
| break; |
| case ES_16: |
| tcg_gen_ld16u_i64(o->out, ptr, 0); |
| break; |
| case ES_32: |
| tcg_gen_ld32u_i64(o->out, ptr, 0); |
| break; |
| case ES_64: |
| tcg_gen_ld_i64(o->out, ptr, 0); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vllez(DisasContext *s, DisasOps *o) |
| { |
| uint8_t es = get_field(s, m3); |
| uint8_t enr; |
| TCGv_i64 t; |
| |
| switch (es) { |
| /* rightmost sub-element of leftmost doubleword */ |
| case ES_8: |
| enr = 7; |
| break; |
| case ES_16: |
| enr = 3; |
| break; |
| case ES_32: |
| enr = 1; |
| break; |
| case ES_64: |
| enr = 0; |
| break; |
| /* leftmost sub-element of leftmost doubleword */ |
| case 6: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| es = ES_32; |
| enr = 0; |
| break; |
| } |
| /* fallthrough */ |
| default: |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| t = tcg_temp_new_i64(); |
| tcg_gen_qemu_ld_i64(t, o->addr1, get_mem_index(s), MO_TE | es); |
| gen_gvec_dup_imm(es, get_field(s, v1), 0); |
| write_vec_element_i64(t, get_field(s, v1), enr, es); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlm(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t v3 = get_field(s, v3); |
| uint8_t v1 = get_field(s, v1); |
| TCGv_i64 t0, t1; |
| |
| if (v3 < v1 || (v3 - v1 + 1) > 16) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| /* |
| * Check for possible access exceptions by trying to load the last |
| * element. The first element will be checked first next. |
| */ |
| t0 = tcg_temp_new_i64(); |
| t1 = tcg_temp_new_i64(); |
| gen_addi_and_wrap_i64(s, t0, o->addr1, (v3 - v1) * 16 + 8); |
| tcg_gen_qemu_ld_i64(t0, t0, get_mem_index(s), MO_TEUQ); |
| |
| for (;; v1++) { |
| tcg_gen_qemu_ld_i64(t1, o->addr1, get_mem_index(s), MO_TEUQ); |
| write_vec_element_i64(t1, v1, 0, ES_64); |
| if (v1 == v3) { |
| break; |
| } |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| tcg_gen_qemu_ld_i64(t1, o->addr1, get_mem_index(s), MO_TEUQ); |
| write_vec_element_i64(t1, v1, 1, ES_64); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| } |
| |
| /* Store the last element, loaded first */ |
| write_vec_element_i64(t0, v1, 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlbb(DisasContext *s, DisasOps *o) |
| { |
| const int64_t block_size = (1ull << (get_field(s, m3) + 6)); |
| const int v1_offs = vec_full_reg_offset(get_field(s, v1)); |
| TCGv_ptr a0; |
| TCGv_i64 bytes; |
| |
| if (get_field(s, m3) > 6) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| bytes = tcg_temp_new_i64(); |
| a0 = tcg_temp_new_ptr(); |
| /* calculate the number of bytes until the next block boundary */ |
| tcg_gen_ori_i64(bytes, o->addr1, -block_size); |
| tcg_gen_neg_i64(bytes, bytes); |
| |
| tcg_gen_addi_ptr(a0, cpu_env, v1_offs); |
| gen_helper_vll(cpu_env, a0, o->addr1, bytes); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlvg(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| TCGv_ptr ptr; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| /* fast path if we don't need the register content */ |
| if (!get_field(s, b2)) { |
| uint8_t enr = get_field(s, d2) & (NUM_VEC_ELEMENTS(es) - 1); |
| |
| write_vec_element_i64(o->in2, get_field(s, v1), enr, es); |
| return DISAS_NEXT; |
| } |
| |
| ptr = tcg_temp_new_ptr(); |
| get_vec_element_ptr_i64(ptr, get_field(s, v1), o->addr1, es); |
| switch (es) { |
| case ES_8: |
| tcg_gen_st8_i64(o->in2, ptr, 0); |
| break; |
| case ES_16: |
| tcg_gen_st16_i64(o->in2, ptr, 0); |
| break; |
| case ES_32: |
| tcg_gen_st32_i64(o->in2, ptr, 0); |
| break; |
| case ES_64: |
| tcg_gen_st_i64(o->in2, ptr, 0); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlvgp(DisasContext *s, DisasOps *o) |
| { |
| write_vec_element_i64(o->in1, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(o->in2, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vll(DisasContext *s, DisasOps *o) |
| { |
| const int v1_offs = vec_full_reg_offset(get_field(s, v1)); |
| TCGv_ptr a0 = tcg_temp_new_ptr(); |
| |
| /* convert highest index into an actual length */ |
| tcg_gen_addi_i64(o->in2, o->in2, 1); |
| tcg_gen_addi_ptr(a0, cpu_env, v1_offs); |
| gen_helper_vll(cpu_env, a0, o->addr1, o->in2); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vmr(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t v1 = get_field(s, v1); |
| const uint8_t v2 = get_field(s, v2); |
| const uint8_t v3 = get_field(s, v3); |
| const uint8_t es = get_field(s, m4); |
| int dst_idx, src_idx; |
| TCGv_i64 tmp; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| if (s->fields.op2 == 0x61) { |
| /* iterate backwards to avoid overwriting data we might need later */ |
| for (dst_idx = NUM_VEC_ELEMENTS(es) - 1; dst_idx >= 0; dst_idx--) { |
| src_idx = dst_idx / 2; |
| if (dst_idx % 2 == 0) { |
| read_vec_element_i64(tmp, v2, src_idx, es); |
| } else { |
| read_vec_element_i64(tmp, v3, src_idx, es); |
| } |
| write_vec_element_i64(tmp, v1, dst_idx, es); |
| } |
| } else { |
| /* iterate forward to avoid overwriting data we might need later */ |
| for (dst_idx = 0; dst_idx < NUM_VEC_ELEMENTS(es); dst_idx++) { |
| src_idx = (dst_idx + NUM_VEC_ELEMENTS(es)) / 2; |
| if (dst_idx % 2 == 0) { |
| read_vec_element_i64(tmp, v2, src_idx, es); |
| } else { |
| read_vec_element_i64(tmp, v3, src_idx, es); |
| } |
| write_vec_element_i64(tmp, v1, dst_idx, es); |
| } |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vpk(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t v1 = get_field(s, v1); |
| const uint8_t v2 = get_field(s, v2); |
| const uint8_t v3 = get_field(s, v3); |
| const uint8_t es = get_field(s, m4); |
| static gen_helper_gvec_3 * const vpk[3] = { |
| gen_helper_gvec_vpk16, |
| gen_helper_gvec_vpk32, |
| gen_helper_gvec_vpk64, |
| }; |
| static gen_helper_gvec_3 * const vpks[3] = { |
| gen_helper_gvec_vpks16, |
| gen_helper_gvec_vpks32, |
| gen_helper_gvec_vpks64, |
| }; |
| static gen_helper_gvec_3_ptr * const vpks_cc[3] = { |
| gen_helper_gvec_vpks_cc16, |
| gen_helper_gvec_vpks_cc32, |
| gen_helper_gvec_vpks_cc64, |
| }; |
| static gen_helper_gvec_3 * const vpkls[3] = { |
| gen_helper_gvec_vpkls16, |
| gen_helper_gvec_vpkls32, |
| gen_helper_gvec_vpkls64, |
| }; |
| static gen_helper_gvec_3_ptr * const vpkls_cc[3] = { |
| gen_helper_gvec_vpkls_cc16, |
| gen_helper_gvec_vpkls_cc32, |
| gen_helper_gvec_vpkls_cc64, |
| }; |
| |
| if (es == ES_8 || es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| switch (s->fields.op2) { |
| case 0x97: |
| if (get_field(s, m5) & 0x1) { |
| gen_gvec_3_ptr(v1, v2, v3, cpu_env, 0, vpks_cc[es - 1]); |
| set_cc_static(s); |
| } else { |
| gen_gvec_3_ool(v1, v2, v3, 0, vpks[es - 1]); |
| } |
| break; |
| case 0x95: |
| if (get_field(s, m5) & 0x1) { |
| gen_gvec_3_ptr(v1, v2, v3, cpu_env, 0, vpkls_cc[es - 1]); |
| set_cc_static(s); |
| } else { |
| gen_gvec_3_ool(v1, v2, v3, 0, vpkls[es - 1]); |
| } |
| break; |
| case 0x94: |
| /* If sources and destination don't overlap -> fast path */ |
| if (v1 != v2 && v1 != v3) { |
| const uint8_t src_es = get_field(s, m4); |
| const uint8_t dst_es = src_es - 1; |
| TCGv_i64 tmp = tcg_temp_new_i64(); |
| int dst_idx, src_idx; |
| |
| for (dst_idx = 0; dst_idx < NUM_VEC_ELEMENTS(dst_es); dst_idx++) { |
| src_idx = dst_idx; |
| if (src_idx < NUM_VEC_ELEMENTS(src_es)) { |
| read_vec_element_i64(tmp, v2, src_idx, src_es); |
| } else { |
| src_idx -= NUM_VEC_ELEMENTS(src_es); |
| read_vec_element_i64(tmp, v3, src_idx, src_es); |
| } |
| write_vec_element_i64(tmp, v1, dst_idx, dst_es); |
| } |
| } else { |
| gen_gvec_3_ool(v1, v2, v3, 0, vpk[es - 1]); |
| } |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vperm(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_4_ool(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), |
| 0, gen_helper_gvec_vperm); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vpdi(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t i2 = extract32(get_field(s, m4), 2, 1); |
| const uint8_t i3 = extract32(get_field(s, m4), 0, 1); |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| |
| read_vec_element_i64(t0, get_field(s, v2), i2, ES_64); |
| read_vec_element_i64(t1, get_field(s, v3), i3, ES_64); |
| write_vec_element_i64(t0, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(t1, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vrep(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t enr = get_field(s, i2); |
| const uint8_t es = get_field(s, m4); |
| |
| if (es > ES_64 || !valid_vec_element(enr, es)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tcg_gen_gvec_dup_mem(es, vec_full_reg_offset(get_field(s, v1)), |
| vec_reg_offset(get_field(s, v3), enr, es), |
| 16, 16); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vrepi(DisasContext *s, DisasOps *o) |
| { |
| const int64_t data = (int16_t)get_field(s, i2); |
| const uint8_t es = get_field(s, m3); |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_dup_imm(es, get_field(s, v1), data); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vsce(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = s->insn->data; |
| const uint8_t enr = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (!valid_vec_element(enr, es)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| read_vec_element_i64(tmp, get_field(s, v2), enr, es); |
| tcg_gen_add_i64(o->addr1, o->addr1, tmp); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 0); |
| |
| read_vec_element_i64(tmp, get_field(s, v1), enr, es); |
| tcg_gen_qemu_st_i64(tmp, o->addr1, get_mem_index(s), MO_TE | es); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vsel(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_4(bitsel, ES_8, get_field(s, v1), |
| get_field(s, v4), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vseg(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| int idx1, idx2; |
| TCGv_i64 tmp; |
| |
| switch (es) { |
| case ES_8: |
| idx1 = 7; |
| idx2 = 15; |
| break; |
| case ES_16: |
| idx1 = 3; |
| idx2 = 7; |
| break; |
| case ES_32: |
| idx1 = 1; |
| idx2 = 3; |
| break; |
| default: |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| read_vec_element_i64(tmp, get_field(s, v2), idx1, es | MO_SIGN); |
| write_vec_element_i64(tmp, get_field(s, v1), 0, ES_64); |
| read_vec_element_i64(tmp, get_field(s, v2), idx2, es | MO_SIGN); |
| write_vec_element_i64(tmp, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vst(DisasContext *s, DisasOps *o) |
| { |
| TCGv_i64 tmp; |
| |
| /* Probe write access before actually modifying memory */ |
| gen_helper_probe_write_access(cpu_env, o->addr1, |
| tcg_constant_i64(16)); |
| |
| tmp = tcg_temp_new_i64(); |
| read_vec_element_i64(tmp, get_field(s, v1), 0, ES_64); |
| tcg_gen_qemu_st_i64(tmp, o->addr1, get_mem_index(s), MO_TEUQ); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| read_vec_element_i64(tmp, get_field(s, v1), 1, ES_64); |
| tcg_gen_qemu_st_i64(tmp, o->addr1, get_mem_index(s), MO_TEUQ); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vstebr(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = s->insn->data; |
| const uint8_t enr = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (!valid_vec_element(enr, es)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| read_vec_element_i64(tmp, get_field(s, v1), enr, es); |
| tcg_gen_qemu_st_i64(tmp, o->addr1, get_mem_index(s), MO_LE | es); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vstbr(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| TCGv_i64 t0, t1; |
| |
| if (es < ES_16 || es > ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| /* Probe write access before actually modifying memory */ |
| gen_helper_probe_write_access(cpu_env, o->addr1, tcg_constant_i64(16)); |
| |
| t0 = tcg_temp_new_i64(); |
| t1 = tcg_temp_new_i64(); |
| |
| |
| if (es == ES_128) { |
| read_vec_element_i64(t1, get_field(s, v1), 0, ES_64); |
| read_vec_element_i64(t0, get_field(s, v1), 1, ES_64); |
| goto write; |
| } |
| |
| read_vec_element_i64(t0, get_field(s, v1), 0, ES_64); |
| read_vec_element_i64(t1, get_field(s, v1), 1, ES_64); |
| |
| /* |
| * For 16 and 32-bit elements, the doubleword bswap below will |
| * reverse the order of the elements. Perform a larger order |
| * swap to put them back into place. For the 128-bit "element", |
| * finish the bswap by swapping the doublewords. |
| */ |
| switch (es) { |
| case MO_16: |
| tcg_gen_hswap_i64(t0, t0); |
| tcg_gen_hswap_i64(t1, t1); |
| break; |
| case MO_32: |
| tcg_gen_wswap_i64(t0, t0); |
| tcg_gen_wswap_i64(t1, t1); |
| break; |
| case MO_64: |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| write: |
| tcg_gen_qemu_st_i64(t0, o->addr1, get_mem_index(s), MO_LEUQ); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| tcg_gen_qemu_st_i64(t1, o->addr1, get_mem_index(s), MO_LEUQ); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vste(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = s->insn->data; |
| const uint8_t enr = get_field(s, m3); |
| TCGv_i64 tmp; |
| |
| if (!valid_vec_element(enr, es)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| read_vec_element_i64(tmp, get_field(s, v1), enr, es); |
| tcg_gen_qemu_st_i64(tmp, o->addr1, get_mem_index(s), MO_TE | es); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vster(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| TCGv_i64 t0, t1; |
| |
| if (es < ES_16 || es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| /* Probe write access before actually modifying memory */ |
| gen_helper_probe_write_access(cpu_env, o->addr1, tcg_constant_i64(16)); |
| |
| /* Begin with the two doublewords swapped... */ |
| t0 = tcg_temp_new_i64(); |
| t1 = tcg_temp_new_i64(); |
| read_vec_element_i64(t1, get_field(s, v1), 0, ES_64); |
| read_vec_element_i64(t0, get_field(s, v1), 1, ES_64); |
| |
| /* ... then swap smaller elements within the doublewords as required. */ |
| switch (es) { |
| case MO_16: |
| tcg_gen_hswap_i64(t1, t1); |
| tcg_gen_hswap_i64(t0, t0); |
| break; |
| case MO_32: |
| tcg_gen_wswap_i64(t1, t1); |
| tcg_gen_wswap_i64(t0, t0); |
| break; |
| case MO_64: |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| tcg_gen_qemu_st_i64(t0, o->addr1, get_mem_index(s), MO_TEUQ); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| tcg_gen_qemu_st_i64(t1, o->addr1, get_mem_index(s), MO_TEUQ); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vstm(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t v3 = get_field(s, v3); |
| uint8_t v1 = get_field(s, v1); |
| TCGv_i64 tmp; |
| |
| while (v3 < v1 || (v3 - v1 + 1) > 16) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| /* Probe write access before actually modifying memory */ |
| gen_helper_probe_write_access(cpu_env, o->addr1, |
| tcg_constant_i64((v3 - v1 + 1) * 16)); |
| |
| tmp = tcg_temp_new_i64(); |
| for (;; v1++) { |
| read_vec_element_i64(tmp, v1, 0, ES_64); |
| tcg_gen_qemu_st_i64(tmp, o->addr1, get_mem_index(s), MO_TEUQ); |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| read_vec_element_i64(tmp, v1, 1, ES_64); |
| tcg_gen_qemu_st_i64(tmp, o->addr1, get_mem_index(s), MO_TEUQ); |
| if (v1 == v3) { |
| break; |
| } |
| gen_addi_and_wrap_i64(s, o->addr1, o->addr1, 8); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vstl(DisasContext *s, DisasOps *o) |
| { |
| const int v1_offs = vec_full_reg_offset(get_field(s, v1)); |
| TCGv_ptr a0 = tcg_temp_new_ptr(); |
| |
| /* convert highest index into an actual length */ |
| tcg_gen_addi_i64(o->in2, o->in2, 1); |
| tcg_gen_addi_ptr(a0, cpu_env, v1_offs); |
| gen_helper_vstl(cpu_env, a0, o->addr1, o->in2); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vup(DisasContext *s, DisasOps *o) |
| { |
| const bool logical = s->fields.op2 == 0xd4 || s->fields.op2 == 0xd5; |
| const uint8_t v1 = get_field(s, v1); |
| const uint8_t v2 = get_field(s, v2); |
| const uint8_t src_es = get_field(s, m3); |
| const uint8_t dst_es = src_es + 1; |
| int dst_idx, src_idx; |
| TCGv_i64 tmp; |
| |
| if (src_es > ES_32) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tmp = tcg_temp_new_i64(); |
| if (s->fields.op2 == 0xd7 || s->fields.op2 == 0xd5) { |
| /* iterate backwards to avoid overwriting data we might need later */ |
| for (dst_idx = NUM_VEC_ELEMENTS(dst_es) - 1; dst_idx >= 0; dst_idx--) { |
| src_idx = dst_idx; |
| read_vec_element_i64(tmp, v2, src_idx, |
| src_es | (logical ? 0 : MO_SIGN)); |
| write_vec_element_i64(tmp, v1, dst_idx, dst_es); |
| } |
| |
| } else { |
| /* iterate forward to avoid overwriting data we might need later */ |
| for (dst_idx = 0; dst_idx < NUM_VEC_ELEMENTS(dst_es); dst_idx++) { |
| src_idx = dst_idx + NUM_VEC_ELEMENTS(src_es) / 2; |
| read_vec_element_i64(tmp, v2, src_idx, |
| src_es | (logical ? 0 : MO_SIGN)); |
| write_vec_element_i64(tmp, v1, dst_idx, dst_es); |
| } |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_va(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| |
| if (es > ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } else if (es == ES_128) { |
| gen_gvec128_3_i64(tcg_gen_add2_i64, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| gen_gvec_fn_3(add, es, get_field(s, v1), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_acc(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b, uint8_t es) |
| { |
| const uint8_t msb_bit_nr = NUM_VEC_ELEMENT_BITS(es) - 1; |
| TCGv_i64 msb_mask = tcg_constant_i64(dup_const(es, 1ull << msb_bit_nr)); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| TCGv_i64 t2 = tcg_temp_new_i64(); |
| TCGv_i64 t3 = tcg_temp_new_i64(); |
| |
| /* Calculate the carry into the MSB, ignoring the old MSBs */ |
| tcg_gen_andc_i64(t1, a, msb_mask); |
| tcg_gen_andc_i64(t2, b, msb_mask); |
| tcg_gen_add_i64(t1, t1, t2); |
| /* Calculate the MSB without any carry into it */ |
| tcg_gen_xor_i64(t3, a, b); |
| /* Calculate the carry out of the MSB in the MSB bit position */ |
| tcg_gen_and_i64(d, a, b); |
| tcg_gen_and_i64(t1, t1, t3); |
| tcg_gen_or_i64(d, d, t1); |
| /* Isolate and shift the carry into position */ |
| tcg_gen_and_i64(d, d, msb_mask); |
| tcg_gen_shri_i64(d, d, msb_bit_nr); |
| } |
| |
| static void gen_acc8_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b) |
| { |
| gen_acc(d, a, b, ES_8); |
| } |
| |
| static void gen_acc16_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b) |
| { |
| gen_acc(d, a, b, ES_16); |
| } |
| |
| static void gen_acc_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) |
| { |
| TCGv_i32 t = tcg_temp_new_i32(); |
| |
| tcg_gen_add_i32(t, a, b); |
| tcg_gen_setcond_i32(TCG_COND_LTU, d, t, b); |
| } |
| |
| static void gen_acc_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b) |
| { |
| TCGv_i64 t = tcg_temp_new_i64(); |
| |
| tcg_gen_add_i64(t, a, b); |
| tcg_gen_setcond_i64(TCG_COND_LTU, d, t, b); |
| } |
| |
| static void gen_acc2_i64(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, |
| TCGv_i64 ah, TCGv_i64 bl, TCGv_i64 bh) |
| { |
| TCGv_i64 th = tcg_temp_new_i64(); |
| TCGv_i64 tl = tcg_temp_new_i64(); |
| TCGv_i64 zero = tcg_constant_i64(0); |
| |
| tcg_gen_add2_i64(tl, th, al, zero, bl, zero); |
| tcg_gen_add2_i64(tl, th, th, zero, ah, zero); |
| tcg_gen_add2_i64(tl, dl, tl, th, bh, zero); |
| tcg_gen_mov_i64(dh, zero); |
| } |
| |
| static DisasJumpType op_vacc(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| static const GVecGen3 g[4] = { |
| { .fni8 = gen_acc8_i64, }, |
| { .fni8 = gen_acc16_i64, }, |
| { .fni4 = gen_acc_i32, }, |
| { .fni8 = gen_acc_i64, }, |
| }; |
| |
| if (es > ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } else if (es == ES_128) { |
| gen_gvec128_3_i64(gen_acc2_i64, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| gen_gvec_3(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_ac2_i64(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, TCGv_i64 ah, |
| TCGv_i64 bl, TCGv_i64 bh, TCGv_i64 cl, TCGv_i64 ch) |
| { |
| TCGv_i64 tl = tcg_temp_new_i64(); |
| TCGv_i64 zero = tcg_constant_i64(0); |
| |
| /* extract the carry only */ |
| tcg_gen_extract_i64(tl, cl, 0, 1); |
| tcg_gen_add2_i64(dl, dh, al, ah, bl, bh); |
| tcg_gen_add2_i64(dl, dh, dl, dh, tl, zero); |
| } |
| |
| static DisasJumpType op_vac(DisasContext *s, DisasOps *o) |
| { |
| if (get_field(s, m5) != ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec128_4_i64(gen_ac2_i64, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3), |
| get_field(s, v4)); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_accc2_i64(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, TCGv_i64 ah, |
| TCGv_i64 bl, TCGv_i64 bh, TCGv_i64 cl, TCGv_i64 ch) |
| { |
| TCGv_i64 tl = tcg_temp_new_i64(); |
| TCGv_i64 th = tcg_temp_new_i64(); |
| TCGv_i64 zero = tcg_constant_i64(0); |
| |
| tcg_gen_andi_i64(tl, cl, 1); |
| tcg_gen_add2_i64(tl, th, tl, zero, al, zero); |
| tcg_gen_add2_i64(tl, th, tl, th, bl, zero); |
| tcg_gen_add2_i64(tl, th, th, zero, ah, zero); |
| tcg_gen_add2_i64(tl, dl, tl, th, bh, zero); |
| tcg_gen_mov_i64(dh, zero); |
| } |
| |
| static DisasJumpType op_vaccc(DisasContext *s, DisasOps *o) |
| { |
| if (get_field(s, m5) != ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec128_4_i64(gen_accc2_i64, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3), |
| get_field(s, v4)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vn(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_3(and, ES_8, get_field(s, v1), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vnc(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_3(andc, ES_8, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_avg_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) |
| { |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| |
| tcg_gen_ext_i32_i64(t0, a); |
| tcg_gen_ext_i32_i64(t1, b); |
| tcg_gen_add_i64(t0, t0, t1); |
| tcg_gen_addi_i64(t0, t0, 1); |
| tcg_gen_shri_i64(t0, t0, 1); |
| tcg_gen_extrl_i64_i32(d, t0); |
| } |
| |
| static void gen_avg_i64(TCGv_i64 dl, TCGv_i64 al, TCGv_i64 bl) |
| { |
| TCGv_i64 dh = tcg_temp_new_i64(); |
| TCGv_i64 ah = tcg_temp_new_i64(); |
| TCGv_i64 bh = tcg_temp_new_i64(); |
| |
| /* extending the sign by one bit is sufficient */ |
| tcg_gen_extract_i64(ah, al, 63, 1); |
| tcg_gen_extract_i64(bh, bl, 63, 1); |
| tcg_gen_add2_i64(dl, dh, al, ah, bl, bh); |
| gen_addi2_i64(dl, dh, dl, dh, 1); |
| tcg_gen_extract2_i64(dl, dl, dh, 1); |
| } |
| |
| static DisasJumpType op_vavg(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| static const GVecGen3 g[4] = { |
| { .fno = gen_helper_gvec_vavg8, }, |
| { .fno = gen_helper_gvec_vavg16, }, |
| { .fni4 = gen_avg_i32, }, |
| { .fni8 = gen_avg_i64, }, |
| }; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| gen_gvec_3(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_avgl_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) |
| { |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| |
| tcg_gen_extu_i32_i64(t0, a); |
| tcg_gen_extu_i32_i64(t1, b); |
| tcg_gen_add_i64(t0, t0, t1); |
| tcg_gen_addi_i64(t0, t0, 1); |
| tcg_gen_shri_i64(t0, t0, 1); |
| tcg_gen_extrl_i64_i32(d, t0); |
| } |
| |
| static void gen_avgl_i64(TCGv_i64 dl, TCGv_i64 al, TCGv_i64 bl) |
| { |
| TCGv_i64 dh = tcg_temp_new_i64(); |
| TCGv_i64 zero = tcg_constant_i64(0); |
| |
| tcg_gen_add2_i64(dl, dh, al, zero, bl, zero); |
| gen_addi2_i64(dl, dh, dl, dh, 1); |
| tcg_gen_extract2_i64(dl, dl, dh, 1); |
| } |
| |
| static DisasJumpType op_vavgl(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| static const GVecGen3 g[4] = { |
| { .fno = gen_helper_gvec_vavgl8, }, |
| { .fno = gen_helper_gvec_vavgl16, }, |
| { .fni4 = gen_avgl_i32, }, |
| { .fni8 = gen_avgl_i64, }, |
| }; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| gen_gvec_3(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vcksm(DisasContext *s, DisasOps *o) |
| { |
| TCGv_i32 tmp = tcg_temp_new_i32(); |
| TCGv_i32 sum = tcg_temp_new_i32(); |
| int i; |
| |
| read_vec_element_i32(sum, get_field(s, v3), 1, ES_32); |
| for (i = 0; i < 4; i++) { |
| read_vec_element_i32(tmp, get_field(s, v2), i, ES_32); |
| tcg_gen_add2_i32(tmp, sum, sum, sum, tmp, tmp); |
| } |
| gen_gvec_dup_imm(ES_32, get_field(s, v1), 0); |
| write_vec_element_i32(sum, get_field(s, v1), 1, ES_32); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vec(DisasContext *s, DisasOps *o) |
| { |
| uint8_t es = get_field(s, m3); |
| const uint8_t enr = NUM_VEC_ELEMENTS(es) / 2 - 1; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| if (s->fields.op2 == 0xdb) { |
| es |= MO_SIGN; |
| } |
| |
| o->in1 = tcg_temp_new_i64(); |
| o->in2 = tcg_temp_new_i64(); |
| read_vec_element_i64(o->in1, get_field(s, v1), enr, es); |
| read_vec_element_i64(o->in2, get_field(s, v2), enr, es); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vc(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| TCGCond cond = s->insn->data; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| tcg_gen_gvec_cmp(cond, es, |
| vec_full_reg_offset(get_field(s, v1)), |
| vec_full_reg_offset(get_field(s, v2)), |
| vec_full_reg_offset(get_field(s, v3)), 16, 16); |
| if (get_field(s, m5) & 0x1) { |
| TCGv_i64 low = tcg_temp_new_i64(); |
| TCGv_i64 high = tcg_temp_new_i64(); |
| |
| read_vec_element_i64(high, get_field(s, v1), 0, ES_64); |
| read_vec_element_i64(low, get_field(s, v1), 1, ES_64); |
| gen_op_update2_cc_i64(s, CC_OP_VC, low, high); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static void gen_clz_i32(TCGv_i32 d, TCGv_i32 a) |
| { |
| tcg_gen_clzi_i32(d, a, 32); |
| } |
| |
| static void gen_clz_i64(TCGv_i64 d, TCGv_i64 a) |
| { |
| tcg_gen_clzi_i64(d, a, 64); |
| } |
| |
| static DisasJumpType op_vclz(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| static const GVecGen2 g[4] = { |
| { .fno = gen_helper_gvec_vclz8, }, |
| { .fno = gen_helper_gvec_vclz16, }, |
| { .fni4 = gen_clz_i32, }, |
| { .fni8 = gen_clz_i64, }, |
| }; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| gen_gvec_2(get_field(s, v1), get_field(s, v2), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_ctz_i32(TCGv_i32 d, TCGv_i32 a) |
| { |
| tcg_gen_ctzi_i32(d, a, 32); |
| } |
| |
| static void gen_ctz_i64(TCGv_i64 d, TCGv_i64 a) |
| { |
| tcg_gen_ctzi_i64(d, a, 64); |
| } |
| |
| static DisasJumpType op_vctz(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| static const GVecGen2 g[4] = { |
| { .fno = gen_helper_gvec_vctz8, }, |
| { .fno = gen_helper_gvec_vctz16, }, |
| { .fni4 = gen_ctz_i32, }, |
| { .fni8 = gen_ctz_i64, }, |
| }; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| gen_gvec_2(get_field(s, v1), get_field(s, v2), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vx(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_3(xor, ES_8, get_field(s, v1), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vgfm(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| static const GVecGen3 g[4] = { |
| { .fno = gen_helper_gvec_vgfm8, }, |
| { .fno = gen_helper_gvec_vgfm16, }, |
| { .fno = gen_helper_gvec_vgfm32, }, |
| { .fno = gen_helper_gvec_vgfm64, }, |
| }; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| gen_gvec_3(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vgfma(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m5); |
| static const GVecGen4 g[4] = { |
| { .fno = gen_helper_gvec_vgfma8, }, |
| { .fno = gen_helper_gvec_vgfma16, }, |
| { .fno = gen_helper_gvec_vgfma32, }, |
| { .fno = gen_helper_gvec_vgfma64, }, |
| }; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| gen_gvec_4(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlc(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_fn_2(neg, es, get_field(s, v1), get_field(s, v2)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vlp(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_fn_2(abs, es, get_field(s, v1), get_field(s, v2)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vmx(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t v1 = get_field(s, v1); |
| const uint8_t v2 = get_field(s, v2); |
| const uint8_t v3 = get_field(s, v3); |
| const uint8_t es = get_field(s, m4); |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| switch (s->fields.op2) { |
| case 0xff: |
| gen_gvec_fn_3(smax, es, v1, v2, v3); |
| break; |
| case 0xfd: |
| gen_gvec_fn_3(umax, es, v1, v2, v3); |
| break; |
| case 0xfe: |
| gen_gvec_fn_3(smin, es, v1, v2, v3); |
| break; |
| case 0xfc: |
| gen_gvec_fn_3(umin, es, v1, v2, v3); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static void gen_mal_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b, TCGv_i32 c) |
| { |
| TCGv_i32 t0 = tcg_temp_new_i32(); |
| |
| tcg_gen_mul_i32(t0, a, b); |
| tcg_gen_add_i32(d, t0, c); |
| } |
| |
| static void gen_mah_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b, TCGv_i32 c) |
| { |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| TCGv_i64 t2 = tcg_temp_new_i64(); |
| |
| tcg_gen_ext_i32_i64(t0, a); |
| tcg_gen_ext_i32_i64(t1, b); |
| tcg_gen_ext_i32_i64(t2, c); |
| tcg_gen_mul_i64(t0, t0, t1); |
| tcg_gen_add_i64(t0, t0, t2); |
| tcg_gen_extrh_i64_i32(d, t0); |
| } |
| |
| static void gen_malh_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b, TCGv_i32 c) |
| { |
| TCGv_i64 t0 = tcg_temp_new_i64(); |
| TCGv_i64 t1 = tcg_temp_new_i64(); |
| TCGv_i64 t2 = tcg_temp_new_i64(); |
| |
| tcg_gen_extu_i32_i64(t0, a); |
| tcg_gen_extu_i32_i64(t1, b); |
| tcg_gen_extu_i32_i64(t2, c); |
| tcg_gen_mul_i64(t0, t0, t1); |
| tcg_gen_add_i64(t0, t0, t2); |
| tcg_gen_extrh_i64_i32(d, t0); |
| } |
| |
| static DisasJumpType op_vma(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m5); |
| static const GVecGen4 g_vmal[3] = { |
| { .fno = gen_helper_gvec_vmal8, }, |
| { .fno = gen_helper_gvec_vmal16, }, |
| { .fni4 = gen_mal_i32, }, |
| }; |
| static const GVecGen4 g_vmah[3] = { |
| { .fno = gen_helper_gvec_vmah8, }, |
| { .fno = gen_helper_gvec_vmah16, }, |
| { .fni4 = gen_mah_i32, }, |
| }; |
| static const GVecGen4 g_vmalh[3] = { |
| { .fno = gen_helper_gvec_vmalh8, }, |
| { .fno = gen_helper_gvec_vmalh16, }, |
| { .fni4 = gen_malh_i32, }, |
| }; |
| static const GVecGen4 g_vmae[3] = { |
| { .fno = gen_helper_gvec_vmae8, }, |
| { .fno = gen_helper_gvec_vmae16, }, |
| { .fno = gen_helper_gvec_vmae32, }, |
| }; |
| static const GVecGen4 g_vmale[3] = { |
| { .fno = gen_helper_gvec_vmale8, }, |
| { .fno = gen_helper_gvec_vmale16, }, |
| { .fno = gen_helper_gvec_vmale32, }, |
| }; |
| static const GVecGen4 g_vmao[3] = { |
| { .fno = gen_helper_gvec_vmao8, }, |
| { .fno = gen_helper_gvec_vmao16, }, |
| { .fno = gen_helper_gvec_vmao32, }, |
| }; |
| static const GVecGen4 g_vmalo[3] = { |
| { .fno = gen_helper_gvec_vmalo8, }, |
| { .fno = gen_helper_gvec_vmalo16, }, |
| { .fno = gen_helper_gvec_vmalo32, }, |
| }; |
| const GVecGen4 *fn; |
| |
| if (es > ES_32) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| switch (s->fields.op2) { |
| case 0xaa: |
| fn = &g_vmal[es]; |
| break; |
| case 0xab: |
| fn = &g_vmah[es]; |
| break; |
| case 0xa9: |
| fn = &g_vmalh[es]; |
| break; |
| case 0xae: |
| fn = &g_vmae[es]; |
| break; |
| case 0xac: |
| fn = &g_vmale[es]; |
| break; |
| case 0xaf: |
| fn = &g_vmao[es]; |
| break; |
| case 0xad: |
| fn = &g_vmalo[es]; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| gen_gvec_4(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), fn); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_mh_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) |
| { |
| TCGv_i32 t = tcg_temp_new_i32(); |
| |
| tcg_gen_muls2_i32(t, d, a, b); |
| } |
| |
| static void gen_mlh_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) |
| { |
| TCGv_i32 t = tcg_temp_new_i32(); |
| |
| tcg_gen_mulu2_i32(t, d, a, b); |
| } |
| |
| static DisasJumpType op_vm(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| static const GVecGen3 g_vmh[3] = { |
| { .fno = gen_helper_gvec_vmh8, }, |
| { .fno = gen_helper_gvec_vmh16, }, |
| { .fni4 = gen_mh_i32, }, |
| }; |
| static const GVecGen3 g_vmlh[3] = { |
| { .fno = gen_helper_gvec_vmlh8, }, |
| { .fno = gen_helper_gvec_vmlh16, }, |
| { .fni4 = gen_mlh_i32, }, |
| }; |
| static const GVecGen3 g_vme[3] = { |
| { .fno = gen_helper_gvec_vme8, }, |
| { .fno = gen_helper_gvec_vme16, }, |
| { .fno = gen_helper_gvec_vme32, }, |
| }; |
| static const GVecGen3 g_vmle[3] = { |
| { .fno = gen_helper_gvec_vmle8, }, |
| { .fno = gen_helper_gvec_vmle16, }, |
| { .fno = gen_helper_gvec_vmle32, }, |
| }; |
| static const GVecGen3 g_vmo[3] = { |
| { .fno = gen_helper_gvec_vmo8, }, |
| { .fno = gen_helper_gvec_vmo16, }, |
| { .fno = gen_helper_gvec_vmo32, }, |
| }; |
| static const GVecGen3 g_vmlo[3] = { |
| { .fno = gen_helper_gvec_vmlo8, }, |
| { .fno = gen_helper_gvec_vmlo16, }, |
| { .fno = gen_helper_gvec_vmlo32, }, |
| }; |
| const GVecGen3 *fn; |
| |
| if (es > ES_32) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| switch (s->fields.op2) { |
| case 0xa2: |
| gen_gvec_fn_3(mul, es, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3)); |
| return DISAS_NEXT; |
| case 0xa3: |
| fn = &g_vmh[es]; |
| break; |
| case 0xa1: |
| fn = &g_vmlh[es]; |
| break; |
| case 0xa6: |
| fn = &g_vme[es]; |
| break; |
| case 0xa4: |
| fn = &g_vmle[es]; |
| break; |
| case 0xa7: |
| fn = &g_vmo[es]; |
| break; |
| case 0xa5: |
| fn = &g_vmlo[es]; |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| gen_gvec_3(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), fn); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vmsl(DisasContext *s, DisasOps *o) |
| { |
| TCGv_i64 l1, h1, l2, h2; |
| |
| if (get_field(s, m5) != ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| l1 = tcg_temp_new_i64(); |
| h1 = tcg_temp_new_i64(); |
| l2 = tcg_temp_new_i64(); |
| h2 = tcg_temp_new_i64(); |
| |
| /* Multiply both even elements from v2 and v3 */ |
| read_vec_element_i64(l1, get_field(s, v2), 0, ES_64); |
| read_vec_element_i64(h1, get_field(s, v3), 0, ES_64); |
| tcg_gen_mulu2_i64(l1, h1, l1, h1); |
| /* Shift result left by one (x2) if requested */ |
| if (extract32(get_field(s, m6), 3, 1)) { |
| tcg_gen_add2_i64(l1, h1, l1, h1, l1, h1); |
| } |
| |
| /* Multiply both odd elements from v2 and v3 */ |
| read_vec_element_i64(l2, get_field(s, v2), 1, ES_64); |
| read_vec_element_i64(h2, get_field(s, v3), 1, ES_64); |
| tcg_gen_mulu2_i64(l2, h2, l2, h2); |
| /* Shift result left by one (x2) if requested */ |
| if (extract32(get_field(s, m6), 2, 1)) { |
| tcg_gen_add2_i64(l2, h2, l2, h2, l2, h2); |
| } |
| |
| /* Add both intermediate results */ |
| tcg_gen_add2_i64(l1, h1, l1, h1, l2, h2); |
| /* Add whole v4 */ |
| read_vec_element_i64(h2, get_field(s, v4), 0, ES_64); |
| read_vec_element_i64(l2, get_field(s, v4), 1, ES_64); |
| tcg_gen_add2_i64(l1, h1, l1, h1, l2, h2); |
| |
| /* Store final result into v1. */ |
| write_vec_element_i64(h1, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(l1, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vnn(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_3(nand, ES_8, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vno(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_3(nor, ES_8, get_field(s, v1), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vnx(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_3(eqv, ES_8, get_field(s, v1), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vo(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_3(or, ES_8, get_field(s, v1), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_voc(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_fn_3(orc, ES_8, get_field(s, v1), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vpopct(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| static const GVecGen2 g[4] = { |
| { .fno = gen_helper_gvec_vpopct8, }, |
| { .fno = gen_helper_gvec_vpopct16, }, |
| { .fni4 = tcg_gen_ctpop_i32, }, |
| { .fni8 = tcg_gen_ctpop_i64, }, |
| }; |
| |
| if (es > ES_64 || (es != ES_8 && !s390_has_feat(S390_FEAT_VECTOR_ENH))) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_2(get_field(s, v1), get_field(s, v2), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_rim_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b, int32_t c) |
| { |
| TCGv_i32 t = tcg_temp_new_i32(); |
| |
| tcg_gen_rotli_i32(t, a, c & 31); |
| tcg_gen_and_i32(t, t, b); |
| tcg_gen_andc_i32(d, d, b); |
| tcg_gen_or_i32(d, d, t); |
| } |
| |
| static void gen_rim_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b, int64_t c) |
| { |
| TCGv_i64 t = tcg_temp_new_i64(); |
| |
| tcg_gen_rotli_i64(t, a, c & 63); |
| tcg_gen_and_i64(t, t, b); |
| tcg_gen_andc_i64(d, d, b); |
| tcg_gen_or_i64(d, d, t); |
| } |
| |
| static DisasJumpType op_verim(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m5); |
| const uint8_t i4 = get_field(s, i4) & |
| (NUM_VEC_ELEMENT_BITS(es) - 1); |
| static const GVecGen3i g[4] = { |
| { .fno = gen_helper_gvec_verim8, }, |
| { .fno = gen_helper_gvec_verim16, }, |
| { .fni4 = gen_rim_i32, |
| .load_dest = true, }, |
| { .fni8 = gen_rim_i64, |
| .load_dest = true, }, |
| }; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_3i(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), i4, &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vesv(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| const uint8_t v1 = get_field(s, v1); |
| const uint8_t v2 = get_field(s, v2); |
| const uint8_t v3 = get_field(s, v3); |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| switch (s->fields.op2) { |
| case 0x70: |
| gen_gvec_fn_3(shlv, es, v1, v2, v3); |
| break; |
| case 0x73: |
| gen_gvec_fn_3(rotlv, es, v1, v2, v3); |
| break; |
| case 0x7a: |
| gen_gvec_fn_3(sarv, es, v1, v2, v3); |
| break; |
| case 0x78: |
| gen_gvec_fn_3(shrv, es, v1, v2, v3); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_ves(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| const uint8_t d2 = get_field(s, d2) & |
| (NUM_VEC_ELEMENT_BITS(es) - 1); |
| const uint8_t v1 = get_field(s, v1); |
| const uint8_t v3 = get_field(s, v3); |
| TCGv_i32 shift; |
| |
| if (es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| if (likely(!get_field(s, b2))) { |
| switch (s->fields.op2) { |
| case 0x30: |
| gen_gvec_fn_2i(shli, es, v1, v3, d2); |
| break; |
| case 0x33: |
| gen_gvec_fn_2i(rotli, es, v1, v3, d2); |
| break; |
| case 0x3a: |
| gen_gvec_fn_2i(sari, es, v1, v3, d2); |
| break; |
| case 0x38: |
| gen_gvec_fn_2i(shri, es, v1, v3, d2); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } else { |
| shift = tcg_temp_new_i32(); |
| tcg_gen_extrl_i64_i32(shift, o->addr1); |
| tcg_gen_andi_i32(shift, shift, NUM_VEC_ELEMENT_BITS(es) - 1); |
| switch (s->fields.op2) { |
| case 0x30: |
| gen_gvec_fn_2s(shls, es, v1, v3, shift); |
| break; |
| case 0x33: |
| gen_gvec_fn_2s(rotls, es, v1, v3, shift); |
| break; |
| case 0x3a: |
| gen_gvec_fn_2s(sars, es, v1, v3, shift); |
| break; |
| case 0x38: |
| gen_gvec_fn_2s(shrs, es, v1, v3, shift); |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType gen_vsh_by_byte(DisasContext *s, DisasOps *o, |
| gen_helper_gvec_2i *gen, |
| gen_helper_gvec_3 *gen_ve2) |
| { |
| bool byte = s->insn->data; |
| |
| if (!byte && s390_has_feat(S390_FEAT_VECTOR_ENH2)) { |
| gen_gvec_3_ool(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), 0, gen_ve2); |
| } else { |
| TCGv_i64 shift = tcg_temp_new_i64(); |
| |
| read_vec_element_i64(shift, get_field(s, v3), 7, ES_8); |
| tcg_gen_andi_i64(shift, shift, byte ? 0x78 : 7); |
| gen_gvec_2i_ool(get_field(s, v1), get_field(s, v2), shift, 0, gen); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vsl(DisasContext *s, DisasOps *o) |
| { |
| return gen_vsh_by_byte(s, o, gen_helper_gvec_vsl, |
| gen_helper_gvec_vsl_ve2); |
| } |
| |
| static DisasJumpType op_vsra(DisasContext *s, DisasOps *o) |
| { |
| return gen_vsh_by_byte(s, o, gen_helper_gvec_vsra, |
| gen_helper_gvec_vsra_ve2); |
| } |
| |
| static DisasJumpType op_vsrl(DisasContext *s, DisasOps *o) |
| { |
| return gen_vsh_by_byte(s, o, gen_helper_gvec_vsrl, |
| gen_helper_gvec_vsrl_ve2); |
| } |
| |
| static DisasJumpType op_vsld(DisasContext *s, DisasOps *o) |
| { |
| const bool byte = s->insn->data; |
| const uint8_t mask = byte ? 15 : 7; |
| const uint8_t mul = byte ? 8 : 1; |
| const uint8_t i4 = get_field(s, i4); |
| const int right_shift = 64 - (i4 & 7) * mul; |
| TCGv_i64 t0, t1, t2; |
| |
| if (i4 & ~mask) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| t0 = tcg_temp_new_i64(); |
| t1 = tcg_temp_new_i64(); |
| t2 = tcg_temp_new_i64(); |
| |
| if ((i4 & 8) == 0) { |
| read_vec_element_i64(t0, get_field(s, v2), 0, ES_64); |
| read_vec_element_i64(t1, get_field(s, v2), 1, ES_64); |
| read_vec_element_i64(t2, get_field(s, v3), 0, ES_64); |
| } else { |
| read_vec_element_i64(t0, get_field(s, v2), 1, ES_64); |
| read_vec_element_i64(t1, get_field(s, v3), 0, ES_64); |
| read_vec_element_i64(t2, get_field(s, v3), 1, ES_64); |
| } |
| |
| tcg_gen_extract2_i64(t0, t1, t0, right_shift); |
| tcg_gen_extract2_i64(t1, t2, t1, right_shift); |
| |
| write_vec_element_i64(t0, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(t1, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vsrd(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t i4 = get_field(s, i4); |
| TCGv_i64 t0, t1, t2; |
| |
| if (i4 & ~7) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| t0 = tcg_temp_new_i64(); |
| t1 = tcg_temp_new_i64(); |
| t2 = tcg_temp_new_i64(); |
| |
| read_vec_element_i64(t0, get_field(s, v2), 1, ES_64); |
| read_vec_element_i64(t1, get_field(s, v3), 0, ES_64); |
| read_vec_element_i64(t2, get_field(s, v3), 1, ES_64); |
| |
| tcg_gen_extract2_i64(t0, t1, t0, i4); |
| tcg_gen_extract2_i64(t1, t2, t1, i4); |
| |
| write_vec_element_i64(t0, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(t1, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vs(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| |
| if (es > ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } else if (es == ES_128) { |
| gen_gvec128_3_i64(tcg_gen_sub2_i64, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| gen_gvec_fn_3(sub, es, get_field(s, v1), get_field(s, v2), |
| get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_scbi_i32(TCGv_i32 d, TCGv_i32 a, TCGv_i32 b) |
| { |
| tcg_gen_setcond_i32(TCG_COND_GEU, d, a, b); |
| } |
| |
| static void gen_scbi_i64(TCGv_i64 d, TCGv_i64 a, TCGv_i64 b) |
| { |
| tcg_gen_setcond_i64(TCG_COND_GEU, d, a, b); |
| } |
| |
| static void gen_scbi2_i64(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, |
| TCGv_i64 ah, TCGv_i64 bl, TCGv_i64 bh) |
| { |
| TCGv_i64 th = tcg_temp_new_i64(); |
| TCGv_i64 tl = tcg_temp_new_i64(); |
| TCGv_i64 zero = tcg_constant_i64(0); |
| |
| tcg_gen_sub2_i64(tl, th, al, zero, bl, zero); |
| tcg_gen_andi_i64(th, th, 1); |
| tcg_gen_sub2_i64(tl, th, ah, zero, th, zero); |
| tcg_gen_sub2_i64(tl, th, tl, th, bh, zero); |
| /* "invert" the result: -1 -> 0; 0 -> 1 */ |
| tcg_gen_addi_i64(dl, th, 1); |
| tcg_gen_mov_i64(dh, zero); |
| } |
| |
| static DisasJumpType op_vscbi(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| static const GVecGen3 g[4] = { |
| { .fno = gen_helper_gvec_vscbi8, }, |
| { .fno = gen_helper_gvec_vscbi16, }, |
| { .fni4 = gen_scbi_i32, }, |
| { .fni8 = gen_scbi_i64, }, |
| }; |
| |
| if (es > ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } else if (es == ES_128) { |
| gen_gvec128_3_i64(gen_scbi2_i64, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3)); |
| return DISAS_NEXT; |
| } |
| gen_gvec_3(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), &g[es]); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_sbi2_i64(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, TCGv_i64 ah, |
| TCGv_i64 bl, TCGv_i64 bh, TCGv_i64 cl, TCGv_i64 ch) |
| { |
| TCGv_i64 tl = tcg_temp_new_i64(); |
| TCGv_i64 th = tcg_temp_new_i64(); |
| |
| tcg_gen_not_i64(tl, bl); |
| tcg_gen_not_i64(th, bh); |
| gen_ac2_i64(dl, dh, al, ah, tl, th, cl, ch); |
| } |
| |
| static DisasJumpType op_vsbi(DisasContext *s, DisasOps *o) |
| { |
| if (get_field(s, m5) != ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec128_4_i64(gen_sbi2_i64, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3), |
| get_field(s, v4)); |
| return DISAS_NEXT; |
| } |
| |
| static void gen_sbcbi2_i64(TCGv_i64 dl, TCGv_i64 dh, TCGv_i64 al, TCGv_i64 ah, |
| TCGv_i64 bl, TCGv_i64 bh, TCGv_i64 cl, TCGv_i64 ch) |
| { |
| TCGv_i64 th = tcg_temp_new_i64(); |
| TCGv_i64 tl = tcg_temp_new_i64(); |
| |
| tcg_gen_not_i64(tl, bl); |
| tcg_gen_not_i64(th, bh); |
| gen_accc2_i64(dl, dh, al, ah, tl, th, cl, ch); |
| } |
| |
| static DisasJumpType op_vsbcbi(DisasContext *s, DisasOps *o) |
| { |
| if (get_field(s, m5) != ES_128) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec128_4_i64(gen_sbcbi2_i64, get_field(s, v1), |
| get_field(s, v2), get_field(s, v3), |
| get_field(s, v4)); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vsumg(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| TCGv_i64 sum, tmp; |
| uint8_t dst_idx; |
| |
| if (es == ES_8 || es > ES_32) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| sum = tcg_temp_new_i64(); |
| tmp = tcg_temp_new_i64(); |
| for (dst_idx = 0; dst_idx < 2; dst_idx++) { |
| uint8_t idx = dst_idx * NUM_VEC_ELEMENTS(es) / 2; |
| const uint8_t max_idx = idx + NUM_VEC_ELEMENTS(es) / 2 - 1; |
| |
| read_vec_element_i64(sum, get_field(s, v3), max_idx, es); |
| for (; idx <= max_idx; idx++) { |
| read_vec_element_i64(tmp, get_field(s, v2), idx, es); |
| tcg_gen_add_i64(sum, sum, tmp); |
| } |
| write_vec_element_i64(sum, get_field(s, v1), dst_idx, ES_64); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vsumq(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| const uint8_t max_idx = NUM_VEC_ELEMENTS(es) - 1; |
| TCGv_i64 sumh, suml, zero, tmpl; |
| uint8_t idx; |
| |
| if (es < ES_32 || es > ES_64) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| sumh = tcg_temp_new_i64(); |
| suml = tcg_temp_new_i64(); |
| zero = tcg_constant_i64(0); |
| tmpl = tcg_temp_new_i64(); |
| |
| tcg_gen_mov_i64(sumh, zero); |
| read_vec_element_i64(suml, get_field(s, v3), max_idx, es); |
| for (idx = 0; idx <= max_idx; idx++) { |
| read_vec_element_i64(tmpl, get_field(s, v2), idx, es); |
| tcg_gen_add2_i64(suml, sumh, suml, sumh, tmpl, zero); |
| } |
| write_vec_element_i64(sumh, get_field(s, v1), 0, ES_64); |
| write_vec_element_i64(suml, get_field(s, v1), 1, ES_64); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vsum(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| TCGv_i32 sum, tmp; |
| uint8_t dst_idx; |
| |
| if (es > ES_16) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| sum = tcg_temp_new_i32(); |
| tmp = tcg_temp_new_i32(); |
| for (dst_idx = 0; dst_idx < 4; dst_idx++) { |
| uint8_t idx = dst_idx * NUM_VEC_ELEMENTS(es) / 4; |
| const uint8_t max_idx = idx + NUM_VEC_ELEMENTS(es) / 4 - 1; |
| |
| read_vec_element_i32(sum, get_field(s, v3), max_idx, es); |
| for (; idx <= max_idx; idx++) { |
| read_vec_element_i32(tmp, get_field(s, v2), idx, es); |
| tcg_gen_add_i32(sum, sum, tmp); |
| } |
| write_vec_element_i32(sum, get_field(s, v1), dst_idx, ES_32); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vtm(DisasContext *s, DisasOps *o) |
| { |
| gen_gvec_2_ptr(get_field(s, v1), get_field(s, v2), |
| cpu_env, 0, gen_helper_gvec_vtm); |
| set_cc_static(s); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfae(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| const uint8_t m5 = get_field(s, m5); |
| static gen_helper_gvec_3 * const g[3] = { |
| gen_helper_gvec_vfae8, |
| gen_helper_gvec_vfae16, |
| gen_helper_gvec_vfae32, |
| }; |
| static gen_helper_gvec_3_ptr * const g_cc[3] = { |
| gen_helper_gvec_vfae_cc8, |
| gen_helper_gvec_vfae_cc16, |
| gen_helper_gvec_vfae_cc32, |
| }; |
| if (es > ES_32) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| if (extract32(m5, 0, 1)) { |
| gen_gvec_3_ptr(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), cpu_env, m5, g_cc[es]); |
| set_cc_static(s); |
| } else { |
| gen_gvec_3_ool(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), m5, g[es]); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfee(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| const uint8_t m5 = get_field(s, m5); |
| static gen_helper_gvec_3 * const g[3] = { |
| gen_helper_gvec_vfee8, |
| gen_helper_gvec_vfee16, |
| gen_helper_gvec_vfee32, |
| }; |
| static gen_helper_gvec_3_ptr * const g_cc[3] = { |
| gen_helper_gvec_vfee_cc8, |
| gen_helper_gvec_vfee_cc16, |
| gen_helper_gvec_vfee_cc32, |
| }; |
| |
| if (es > ES_32 || m5 & ~0x3) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| if (extract32(m5, 0, 1)) { |
| gen_gvec_3_ptr(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), cpu_env, m5, g_cc[es]); |
| set_cc_static(s); |
| } else { |
| gen_gvec_3_ool(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), m5, g[es]); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfene(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m4); |
| const uint8_t m5 = get_field(s, m5); |
| static gen_helper_gvec_3 * const g[3] = { |
| gen_helper_gvec_vfene8, |
| gen_helper_gvec_vfene16, |
| gen_helper_gvec_vfene32, |
| }; |
| static gen_helper_gvec_3_ptr * const g_cc[3] = { |
| gen_helper_gvec_vfene_cc8, |
| gen_helper_gvec_vfene_cc16, |
| gen_helper_gvec_vfene_cc32, |
| }; |
| |
| if (es > ES_32 || m5 & ~0x3) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| if (extract32(m5, 0, 1)) { |
| gen_gvec_3_ptr(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), cpu_env, m5, g_cc[es]); |
| set_cc_static(s); |
| } else { |
| gen_gvec_3_ool(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), m5, g[es]); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vistr(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m3); |
| const uint8_t m5 = get_field(s, m5); |
| static gen_helper_gvec_2 * const g[3] = { |
| gen_helper_gvec_vistr8, |
| gen_helper_gvec_vistr16, |
| gen_helper_gvec_vistr32, |
| }; |
| static gen_helper_gvec_2_ptr * const g_cc[3] = { |
| gen_helper_gvec_vistr_cc8, |
| gen_helper_gvec_vistr_cc16, |
| gen_helper_gvec_vistr_cc32, |
| }; |
| |
| if (es > ES_32 || m5 & ~0x1) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| if (extract32(m5, 0, 1)) { |
| gen_gvec_2_ptr(get_field(s, v1), get_field(s, v2), |
| cpu_env, 0, g_cc[es]); |
| set_cc_static(s); |
| } else { |
| gen_gvec_2_ool(get_field(s, v1), get_field(s, v2), 0, |
| g[es]); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vstrc(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t es = get_field(s, m5); |
| const uint8_t m6 = get_field(s, m6); |
| static gen_helper_gvec_4 * const g[3] = { |
| gen_helper_gvec_vstrc8, |
| gen_helper_gvec_vstrc16, |
| gen_helper_gvec_vstrc32, |
| }; |
| static gen_helper_gvec_4 * const g_rt[3] = { |
| gen_helper_gvec_vstrc_rt8, |
| gen_helper_gvec_vstrc_rt16, |
| gen_helper_gvec_vstrc_rt32, |
| }; |
| static gen_helper_gvec_4_ptr * const g_cc[3] = { |
| gen_helper_gvec_vstrc_cc8, |
| gen_helper_gvec_vstrc_cc16, |
| gen_helper_gvec_vstrc_cc32, |
| }; |
| static gen_helper_gvec_4_ptr * const g_cc_rt[3] = { |
| gen_helper_gvec_vstrc_cc_rt8, |
| gen_helper_gvec_vstrc_cc_rt16, |
| gen_helper_gvec_vstrc_cc_rt32, |
| }; |
| |
| if (es > ES_32) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| if (extract32(m6, 0, 1)) { |
| if (extract32(m6, 2, 1)) { |
| gen_gvec_4_ptr(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), |
| cpu_env, m6, g_cc_rt[es]); |
| } else { |
| gen_gvec_4_ptr(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), |
| cpu_env, m6, g_cc[es]); |
| } |
| set_cc_static(s); |
| } else { |
| if (extract32(m6, 2, 1)) { |
| gen_gvec_4_ool(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), |
| m6, g_rt[es]); |
| } else { |
| gen_gvec_4_ool(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), |
| m6, g[es]); |
| } |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vstrs(DisasContext *s, DisasOps *o) |
| { |
| typedef void (*helper_vstrs)(TCGv_ptr, TCGv_ptr, TCGv_ptr, |
| TCGv_ptr, TCGv_ptr, TCGv_i32); |
| static const helper_vstrs fns[3][2] = { |
| { gen_helper_gvec_vstrs_8, gen_helper_gvec_vstrs_zs8 }, |
| { gen_helper_gvec_vstrs_16, gen_helper_gvec_vstrs_zs16 }, |
| { gen_helper_gvec_vstrs_32, gen_helper_gvec_vstrs_zs32 }, |
| }; |
| const uint8_t es = get_field(s, m5); |
| const uint8_t m6 = get_field(s, m6); |
| const bool zs = extract32(m6, 1, 1); |
| |
| if (es > ES_32 || m6 & ~2) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_4_ptr(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), |
| cpu_env, 0, fns[es][zs]); |
| set_cc_static(s); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfa(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t fpf = get_field(s, m4); |
| const uint8_t m5 = get_field(s, m5); |
| gen_helper_gvec_3_ptr *fn = NULL; |
| |
| switch (s->fields.op2) { |
| case 0xe3: |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfa32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfa64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfa128; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xe5: |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfd32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfd64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfd128; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xe7: |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfm32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfm64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfm128; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xe2: |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfs32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfs64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfs128; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| if (!fn || extract32(m5, 0, 3)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_3_ptr(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), cpu_env, m5, fn); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_wfc(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t fpf = get_field(s, m3); |
| const uint8_t m4 = get_field(s, m4); |
| gen_helper_gvec_2_ptr *fn = NULL; |
| |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_wfk32; |
| if (s->fields.op2 == 0xcb) { |
| fn = gen_helper_gvec_wfc32; |
| } |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_wfk64; |
| if (s->fields.op2 == 0xcb) { |
| fn = gen_helper_gvec_wfc64; |
| } |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_wfk128; |
| if (s->fields.op2 == 0xcb) { |
| fn = gen_helper_gvec_wfc128; |
| } |
| } |
| break; |
| default: |
| break; |
| }; |
| |
| if (!fn || m4) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_2_ptr(get_field(s, v1), get_field(s, v2), cpu_env, 0, fn); |
| set_cc_static(s); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfc(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t fpf = get_field(s, m4); |
| const uint8_t m5 = get_field(s, m5); |
| const uint8_t m6 = get_field(s, m6); |
| const bool cs = extract32(m6, 0, 1); |
| const bool sq = extract32(m5, 2, 1); |
| gen_helper_gvec_3_ptr *fn = NULL; |
| |
| switch (s->fields.op2) { |
| case 0xe8: |
| switch (fpf) { |
| case FPF_SHORT: |
| fn = cs ? gen_helper_gvec_vfce32_cc : gen_helper_gvec_vfce32; |
| break; |
| case FPF_LONG: |
| fn = cs ? gen_helper_gvec_vfce64_cc : gen_helper_gvec_vfce64; |
| break; |
| case FPF_EXT: |
| fn = cs ? gen_helper_gvec_vfce128_cc : gen_helper_gvec_vfce128; |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xeb: |
| switch (fpf) { |
| case FPF_SHORT: |
| fn = cs ? gen_helper_gvec_vfch32_cc : gen_helper_gvec_vfch32; |
| break; |
| case FPF_LONG: |
| fn = cs ? gen_helper_gvec_vfch64_cc : gen_helper_gvec_vfch64; |
| break; |
| case FPF_EXT: |
| fn = cs ? gen_helper_gvec_vfch128_cc : gen_helper_gvec_vfch128; |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xea: |
| switch (fpf) { |
| case FPF_SHORT: |
| fn = cs ? gen_helper_gvec_vfche32_cc : gen_helper_gvec_vfche32; |
| break; |
| case FPF_LONG: |
| fn = cs ? gen_helper_gvec_vfche64_cc : gen_helper_gvec_vfche64; |
| break; |
| case FPF_EXT: |
| fn = cs ? gen_helper_gvec_vfche128_cc : gen_helper_gvec_vfche128; |
| break; |
| default: |
| break; |
| } |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| if (!fn || extract32(m5, 0, 2) || extract32(m6, 1, 3) || |
| (!s390_has_feat(S390_FEAT_VECTOR_ENH) && (fpf != FPF_LONG || sq))) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_3_ptr(get_field(s, v1), get_field(s, v2), get_field(s, v3), |
| cpu_env, m5, fn); |
| if (cs) { |
| set_cc_static(s); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vcdg(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t fpf = get_field(s, m3); |
| const uint8_t m4 = get_field(s, m4); |
| const uint8_t erm = get_field(s, m5); |
| gen_helper_gvec_2_ptr *fn = NULL; |
| |
| |
| switch (s->fields.op2) { |
| case 0xc3: |
| switch (fpf) { |
| case FPF_LONG: |
| fn = gen_helper_gvec_vcdg64; |
| break; |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH2)) { |
| fn = gen_helper_gvec_vcdg32; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xc1: |
| switch (fpf) { |
| case FPF_LONG: |
| fn = gen_helper_gvec_vcdlg64; |
| break; |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH2)) { |
| fn = gen_helper_gvec_vcdlg32; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xc2: |
| switch (fpf) { |
| case FPF_LONG: |
| fn = gen_helper_gvec_vcgd64; |
| break; |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH2)) { |
| fn = gen_helper_gvec_vcgd32; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xc0: |
| switch (fpf) { |
| case FPF_LONG: |
| fn = gen_helper_gvec_vclgd64; |
| break; |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH2)) { |
| fn = gen_helper_gvec_vclgd32; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xc7: |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfi32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfi64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfi128; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0xc5: |
| switch (fpf) { |
| case FPF_LONG: |
| fn = gen_helper_gvec_vflr64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vflr128; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| if (!fn || extract32(m4, 0, 2) || erm > 7 || erm == 2) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_2_ptr(get_field(s, v1), get_field(s, v2), cpu_env, |
| deposit32(m4, 4, 4, erm), fn); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfll(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t fpf = get_field(s, m3); |
| const uint8_t m4 = get_field(s, m4); |
| gen_helper_gvec_2_ptr *fn = NULL; |
| |
| switch (fpf) { |
| case FPF_SHORT: |
| fn = gen_helper_gvec_vfll32; |
| break; |
| case FPF_LONG: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfll64; |
| } |
| break; |
| default: |
| break; |
| } |
| |
| if (!fn || extract32(m4, 0, 3)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_2_ptr(get_field(s, v1), get_field(s, v2), cpu_env, m4, fn); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfmax(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t fpf = get_field(s, m4); |
| const uint8_t m6 = get_field(s, m6); |
| const uint8_t m5 = get_field(s, m5); |
| gen_helper_gvec_3_ptr *fn; |
| |
| if (m6 == 5 || m6 == 6 || m6 == 7 || m6 > 13) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s->fields.op2 == 0xef) { |
| fn = gen_helper_gvec_vfmax32; |
| } else { |
| fn = gen_helper_gvec_vfmin32; |
| } |
| break; |
| case FPF_LONG: |
| if (s->fields.op2 == 0xef) { |
| fn = gen_helper_gvec_vfmax64; |
| } else { |
| fn = gen_helper_gvec_vfmin64; |
| } |
| break; |
| case FPF_EXT: |
| if (s->fields.op2 == 0xef) { |
| fn = gen_helper_gvec_vfmax128; |
| } else { |
| fn = gen_helper_gvec_vfmin128; |
| } |
| break; |
| default: |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_3_ptr(get_field(s, v1), get_field(s, v2), get_field(s, v3), |
| cpu_env, deposit32(m5, 4, 4, m6), fn); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfma(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t m5 = get_field(s, m5); |
| const uint8_t fpf = get_field(s, m6); |
| gen_helper_gvec_4_ptr *fn = NULL; |
| |
| switch (s->fields.op2) { |
| case 0x8f: |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfma32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfma64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfma128; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0x8e: |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfms32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfms64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfms128; |
| } |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0x9f: |
| switch (fpf) { |
| case FPF_SHORT: |
| fn = gen_helper_gvec_vfnma32; |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfnma64; |
| break; |
| case FPF_EXT: |
| fn = gen_helper_gvec_vfnma128; |
| break; |
| default: |
| break; |
| } |
| break; |
| case 0x9e: |
| switch (fpf) { |
| case FPF_SHORT: |
| fn = gen_helper_gvec_vfnms32; |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfnms64; |
| break; |
| case FPF_EXT: |
| fn = gen_helper_gvec_vfnms128; |
| break; |
| default: |
| break; |
| } |
| break; |
| default: |
| g_assert_not_reached(); |
| } |
| |
| if (!fn || extract32(m5, 0, 3)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_4_ptr(get_field(s, v1), get_field(s, v2), |
| get_field(s, v3), get_field(s, v4), cpu_env, m5, fn); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfpso(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t v1 = get_field(s, v1); |
| const uint8_t v2 = get_field(s, v2); |
| const uint8_t fpf = get_field(s, m3); |
| const uint8_t m4 = get_field(s, m4); |
| const uint8_t m5 = get_field(s, m5); |
| const bool se = extract32(m4, 3, 1); |
| TCGv_i64 tmp; |
| |
| if ((fpf != FPF_LONG && !s390_has_feat(S390_FEAT_VECTOR_ENH)) || |
| extract32(m4, 0, 3) || m5 > 2) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| switch (fpf) { |
| case FPF_SHORT: |
| if (!se) { |
| switch (m5) { |
| case 0: |
| /* sign bit is inverted (complement) */ |
| gen_gvec_fn_2i(xori, ES_32, v1, v2, 1ull << 31); |
| break; |
| case 1: |
| /* sign bit is set to one (negative) */ |
| gen_gvec_fn_2i(ori, ES_32, v1, v2, 1ull << 31); |
| break; |
| case 2: |
| /* sign bit is set to zero (positive) */ |
| gen_gvec_fn_2i(andi, ES_32, v1, v2, (1ull << 31) - 1); |
| break; |
| } |
| return DISAS_NEXT; |
| } |
| break; |
| case FPF_LONG: |
| if (!se) { |
| switch (m5) { |
| case 0: |
| /* sign bit is inverted (complement) */ |
| gen_gvec_fn_2i(xori, ES_64, v1, v2, 1ull << 63); |
| break; |
| case 1: |
| /* sign bit is set to one (negative) */ |
| gen_gvec_fn_2i(ori, ES_64, v1, v2, 1ull << 63); |
| break; |
| case 2: |
| /* sign bit is set to zero (positive) */ |
| gen_gvec_fn_2i(andi, ES_64, v1, v2, (1ull << 63) - 1); |
| break; |
| } |
| return DISAS_NEXT; |
| } |
| break; |
| case FPF_EXT: |
| /* Only a single element. */ |
| break; |
| default: |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| /* With a single element, we are only interested in bit 0. */ |
| tmp = tcg_temp_new_i64(); |
| read_vec_element_i64(tmp, v2, 0, ES_64); |
| switch (m5) { |
| case 0: |
| /* sign bit is inverted (complement) */ |
| tcg_gen_xori_i64(tmp, tmp, 1ull << 63); |
| break; |
| case 1: |
| /* sign bit is set to one (negative) */ |
| tcg_gen_ori_i64(tmp, tmp, 1ull << 63); |
| break; |
| case 2: |
| /* sign bit is set to zero (positive) */ |
| tcg_gen_andi_i64(tmp, tmp, (1ull << 63) - 1); |
| break; |
| } |
| write_vec_element_i64(tmp, v1, 0, ES_64); |
| |
| if (fpf == FPF_EXT) { |
| read_vec_element_i64(tmp, v2, 1, ES_64); |
| write_vec_element_i64(tmp, v1, 1, ES_64); |
| } |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vfsq(DisasContext *s, DisasOps *o) |
| { |
| const uint8_t fpf = get_field(s, m3); |
| const uint8_t m4 = get_field(s, m4); |
| gen_helper_gvec_2_ptr *fn = NULL; |
| |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfsq32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vfsq64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vfsq128; |
| } |
| break; |
| default: |
| break; |
| } |
| |
| if (!fn || extract32(m4, 0, 3)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_2_ptr(get_field(s, v1), get_field(s, v2), cpu_env, m4, fn); |
| return DISAS_NEXT; |
| } |
| |
| static DisasJumpType op_vftci(DisasContext *s, DisasOps *o) |
| { |
| const uint16_t i3 = get_field(s, i3); |
| const uint8_t fpf = get_field(s, m4); |
| const uint8_t m5 = get_field(s, m5); |
| gen_helper_gvec_2_ptr *fn = NULL; |
| |
| switch (fpf) { |
| case FPF_SHORT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vftci32; |
| } |
| break; |
| case FPF_LONG: |
| fn = gen_helper_gvec_vftci64; |
| break; |
| case FPF_EXT: |
| if (s390_has_feat(S390_FEAT_VECTOR_ENH)) { |
| fn = gen_helper_gvec_vftci128; |
| } |
| break; |
| default: |
| break; |
| } |
| |
| if (!fn || extract32(m5, 0, 3)) { |
| gen_program_exception(s, PGM_SPECIFICATION); |
| return DISAS_NORETURN; |
| } |
| |
| gen_gvec_2_ptr(get_field(s, v1), get_field(s, v2), cpu_env, |
| deposit32(m5, 4, 12, i3), fn); |
| set_cc_static(s); |
| return DISAS_NEXT; |
| } |