@@ -1428,3 +1428,12 @@ DEF_HELPER_3(rstsa32, i64, env, i64, i64)
DEF_HELPER_3(urstsa32, i64, env, i64, i64)
DEF_HELPER_3(kstsa32, i64, env, i64, i64)
DEF_HELPER_3(ukstsa32, i64, env, i64, i64)
+
+DEF_HELPER_3(sra32, i64, env, i64, i64)
+DEF_HELPER_3(sra32_u, i64, env, i64, i64)
+DEF_HELPER_3(srl32, i64, env, i64, i64)
+DEF_HELPER_3(srl32_u, i64, env, i64, i64)
+DEF_HELPER_3(sll32, i64, env, i64, i64)
+DEF_HELPER_3(ksll32, i64, env, i64, i64)
+DEF_HELPER_3(kslra32, i64, env, i64, i64)
+DEF_HELPER_3(kslra32_u, i64, env, i64, i64)
@@ -1045,3 +1045,18 @@ rstsa32 1011001 ..... ..... 010 ..... 1110111 @r
urstsa32 1101001 ..... ..... 010 ..... 1110111 @r
kstsa32 1100001 ..... ..... 010 ..... 1110111 @r
ukstsa32 1110001 ..... ..... 010 ..... 1110111 @r
+
+sra32 0101000 ..... ..... 010 ..... 1110111 @r
+sra32_u 0110000 ..... ..... 010 ..... 1110111 @r
+srai32 0111000 ..... ..... 010 ..... 1110111 @sh5
+srai32_u 1000000 ..... ..... 010 ..... 1110111 @sh5
+srl32 0101001 ..... ..... 010 ..... 1110111 @r
+srl32_u 0110001 ..... ..... 010 ..... 1110111 @r
+srli32 0111001 ..... ..... 010 ..... 1110111 @sh5
+srli32_u 1000001 ..... ..... 010 ..... 1110111 @sh5
+sll32 0101010 ..... ..... 010 ..... 1110111 @r
+slli32 0111010 ..... ..... 010 ..... 1110111 @sh5
+ksll32 0110010 ..... ..... 010 ..... 1110111 @r
+kslli32 1000010 ..... ..... 010 ..... 1110111 @sh5
+kslra32 0101011 ..... ..... 010 ..... 1110111 @r
+kslra32_u 0110011 ..... ..... 010 ..... 1110111 @r
@@ -1033,3 +1033,58 @@ GEN_RVP64_R_OOL(rstsa32);
GEN_RVP64_R_OOL(urstsa32);
GEN_RVP64_R_OOL(kstsa32);
GEN_RVP64_R_OOL(ukstsa32);
+
+/* (RV64 Only) SIMD 32-bit Shift Instructions */
+static inline bool
+rvp64_shifti(DisasContext *ctx, arg_shift *a,
+ void (* fn)(TCGv_i64, TCGv_ptr, TCGv_i64, TCGv_i64))
+{
+ TCGv t1;
+ TCGv_i64 src1, dst, shift;
+ if (!has_ext(ctx, RVP)) {
+ return false;
+ }
+
+ src1 = tcg_temp_new_i64();
+ dst = tcg_temp_new_i64();
+ t1 = tcg_temp_new();
+
+ gen_get_gpr(t1, a->rs1);
+ tcg_gen_ext_tl_i64(src1, t1);
+ shift = tcg_const_i64(a->shamt);
+
+ fn(dst, cpu_env, src1, shift);
+ tcg_gen_trunc_i64_tl(t1, dst);
+ gen_set_gpr(a->rd, t1);
+
+ tcg_temp_free_i64(src1);
+ tcg_temp_free_i64(dst);
+ tcg_temp_free_i64(shift);
+ tcg_temp_free(t1);
+ return true;
+}
+
+#define GEN_RVP64_SHIFTI(NAME, OP) \
+static bool trans_##NAME(DisasContext *s, arg_shift *a) \
+{ \
+ REQUIRE_64BIT(s); \
+ return rvp64_shifti(s, a, OP); \
+}
+
+GEN_RVP64_SHIFTI(srai32, gen_helper_sra32);
+GEN_RVP64_SHIFTI(srli32, gen_helper_srl32);
+GEN_RVP64_SHIFTI(slli32, gen_helper_sll32);
+
+GEN_RVP64_SHIFTI(srai32_u, gen_helper_sra32_u);
+GEN_RVP64_SHIFTI(srli32_u, gen_helper_srl32_u);
+GEN_RVP64_SHIFTI(kslli32, gen_helper_ksll32);
+
+GEN_RVP64_R_OOL(sra32);
+GEN_RVP64_R_OOL(srl32);
+GEN_RVP64_R_OOL(sll32);
+GEN_RVP64_R_OOL(ksll32);
+GEN_RVP64_R_OOL(kslra32);
+
+GEN_RVP64_R_OOL(sra32_u);
+GEN_RVP64_R_OOL(srl32_u);
+GEN_RVP64_R_OOL(kslra32_u);
@@ -3263,3 +3263,107 @@ static inline void do_ukstsa32(CPURISCVState *env, void *vd, void *va,
}
RVPR64_64_64(ukstsa32, 2, 4);
+
+/* (RV64 Only) SIMD 32-bit Shift Instructions */
+static inline void do_sra32(CPURISCVState *env, void *vd, void *va,
+ void *vb, uint8_t i)
+{
+ int32_t *d = vd, *a = va;
+ uint8_t shift = *(uint8_t *)vb & 0x1f;
+ d[i] = a[i] >> shift;
+}
+
+RVPR64_64_64(sra32, 1, 4);
+
+static inline void do_srl32(CPURISCVState *env, void *vd, void *va,
+ void *vb, uint8_t i)
+{
+ uint32_t *d = vd, *a = va;
+ uint8_t shift = *(uint8_t *)vb & 0x1f;
+ d[i] = a[i] >> shift;
+}
+
+RVPR64_64_64(srl32, 1, 4);
+
+static inline void do_sll32(CPURISCVState *env, void *vd, void *va,
+ void *vb, uint8_t i)
+{
+ uint32_t *d = vd, *a = va;
+ uint8_t shift = *(uint8_t *)vb & 0x1f;
+ d[i] = a[i] << shift;
+}
+
+RVPR64_64_64(sll32, 1, 4);
+
+static inline void do_sra32_u(CPURISCVState *env, void *vd, void *va,
+ void *vb, uint8_t i)
+{
+ int32_t *d = vd, *a = va;
+ uint8_t shift = *(uint8_t *)vb & 0x1f;
+
+ d[i] = vssra32(env, 0, a[i], shift);
+}
+
+RVPR64_64_64(sra32_u, 1, 4);
+
+static inline void do_srl32_u(CPURISCVState *env, void *vd, void *va,
+ void *vb, uint8_t i)
+{
+ uint32_t *d = vd, *a = va;
+ uint8_t shift = *(uint8_t *)vb & 0x1f;
+
+ d[i] = vssrl32(env, 0, a[i], shift);
+}
+
+RVPR64_64_64(srl32_u, 1, 4);
+
+static inline void do_ksll32(CPURISCVState *env, void *vd, void *va,
+ void *vb, uint8_t i)
+{
+ int32_t *d = vd, *a = va, result;
+ uint8_t shift = *(uint64_t *)vb & 0x1f;
+
+ result = a[i] << shift;
+ if (shift > clrsb32(a[i])) {
+ env->vxsat = 0x1;
+ d[i] = (a[i] & INT32_MIN) ? INT32_MIN : INT32_MAX;
+ } else {
+ d[i] = result;
+ }
+}
+
+RVPR64_64_64(ksll32, 1, 4);
+
+static inline void do_kslra32(CPURISCVState *env, void *vd, void *va,
+ void *vb, uint8_t i)
+{
+ int32_t *d = vd, *a = va;
+ int64_t shift = sextract64(*(uint64_t *)vb, 0, 6);
+
+ if (shift >= 0) {
+ do_ksll32(env, vd, va, vb, i);
+ } else {
+ shift = -shift;
+ shift = (shift == 32) ? 31 : shift;
+ d[i] = a[i] >> shift;
+ }
+}
+
+RVPR64_64_64(kslra32, 1, 4);
+
+static inline void do_kslra32_u(CPURISCVState *env, void *vd, void *va,
+ void *vb, uint8_t i)
+{
+ uint32_t *d = vd, *a = va;
+ int32_t shift = sextract32((*(uint32_t *)vb), 0, 6);
+
+ if (shift >= 0) {
+ do_ksll32(env, vd, va, vb, i);
+ } else {
+ shift = -shift;
+ shift = (shift == 32) ? 31 : shift;
+ d[i] = vssra32(env, 0, a[i], shift);
+ }
+}
+
+RVPR64_64_64(kslra32_u, 1, 4);
SIMD 32-bit right shift with rounding or left shift with saturation. Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com> --- target/riscv/helper.h | 9 ++ target/riscv/insn32.decode | 15 ++++ target/riscv/insn_trans/trans_rvp.c.inc | 55 +++++++++++++ target/riscv/packed_helper.c | 104 ++++++++++++++++++++++++ 4 files changed, 183 insertions(+)