diff mbox series

[5/6] target/riscv: Flush not valid NaN-boxing input to canonical NaN

Message ID 20200626205917.4545-6-zhiwei_liu@c-sky.com (mailing list archive)
State New, archived
Headers show
Series target/riscv: NaN-boxing for multiple precison | expand

Commit Message

LIU Zhiwei June 26, 2020, 8:59 p.m. UTC
Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com>
---
 target/riscv/insn_trans/trans_rvd.inc.c |   7 +-
 target/riscv/insn_trans/trans_rvf.inc.c | 272 ++++++++++++++++++++----
 2 files changed, 235 insertions(+), 44 deletions(-)

Comments

Chih-Min Chao June 30, 2020, 7:31 a.m. UTC | #1
On Sat, Jun 27, 2020 at 5:09 AM LIU Zhiwei <zhiwei_liu@c-sky.com> wrote:

> Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com>
> ---
>  target/riscv/insn_trans/trans_rvd.inc.c |   7 +-
>  target/riscv/insn_trans/trans_rvf.inc.c | 272 ++++++++++++++++++++----
>  2 files changed, 235 insertions(+), 44 deletions(-)
>
> diff --git a/target/riscv/insn_trans/trans_rvd.inc.c
> b/target/riscv/insn_trans/trans_rvd.inc.c
> index c0f4a0c789..16947ea6da 100644
> --- a/target/riscv/insn_trans/trans_rvd.inc.c
> +++ b/target/riscv/insn_trans/trans_rvd.inc.c
> @@ -241,10 +241,15 @@ static bool trans_fcvt_d_s(DisasContext *ctx,
> arg_fcvt_d_s *a)
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVD);
>
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    check_nanboxed(ctx, 1, t1);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fcvt_d_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1]);
> +    gen_helper_fcvt_d_s(cpu_fpr[a->rd], cpu_env, t1);
>
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
>      return true;
>  }
>
> diff --git a/target/riscv/insn_trans/trans_rvf.inc.c
> b/target/riscv/insn_trans/trans_rvf.inc.c
> index 04bc8e5cb5..b0379b9d1f 100644
> --- a/target/riscv/insn_trans/trans_rvf.inc.c
> +++ b/target/riscv/insn_trans/trans_rvf.inc.c
> @@ -58,11 +58,23 @@ static bool trans_fmadd_s(DisasContext *ctx,
> arg_fmadd_s *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    TCGv_i64 t3 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
> +    check_nanboxed(ctx, 3, t1, t2, t3);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fmadd_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
> -                       cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
> +    gen_helper_fmadd_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
> +    tcg_temp_free_i64(t3);
>      return true;
>  }
>
> @@ -70,11 +82,23 @@ static bool trans_fmsub_s(DisasContext *ctx,
> arg_fmsub_s *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    TCGv_i64 t3 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
> +    check_nanboxed(ctx, 3, t1, t2, t3);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fmsub_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
> -                       cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
> +    gen_helper_fmsub_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
> +    tcg_temp_free_i64(t3);
>      return true;
>  }
>
> @@ -82,11 +106,23 @@ static bool trans_fnmsub_s(DisasContext *ctx,
> arg_fnmsub_s *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    TCGv_i64 t3 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
> +    check_nanboxed(ctx, 3, t1, t2, t3);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fnmsub_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
> -                        cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
> +    gen_helper_fnmsub_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
> +    tcg_temp_free_i64(t3);
>      return true;
>  }
>
> @@ -94,11 +130,23 @@ static bool trans_fnmadd_s(DisasContext *ctx,
> arg_fnmadd_s *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    TCGv_i64 t3 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
> +    check_nanboxed(ctx, 3, t1, t2, t3);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fnmadd_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
> -                        cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
> -    mark_fs_dirty(ctx);
> +    gen_helper_fnmadd_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>      gen_nanbox_fpr(ctx, a->rd);
> +
> +    mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
> +    tcg_temp_free_i64(t3);
>      return true;
>  }
>
> @@ -107,11 +155,19 @@ static bool trans_fadd_s(DisasContext *ctx,
> arg_fadd_s *a)
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
>
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fadd_s(cpu_fpr[a->rd], cpu_env,
> -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
> +    gen_helper_fadd_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -120,11 +176,19 @@ static bool trans_fsub_s(DisasContext *ctx,
> arg_fsub_s *a)
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
>
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fsub_s(cpu_fpr[a->rd], cpu_env,
> -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
> +    gen_helper_fsub_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -133,11 +197,19 @@ static bool trans_fmul_s(DisasContext *ctx,
> arg_fmul_s *a)
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
>
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fmul_s(cpu_fpr[a->rd], cpu_env,
> -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
> +    gen_helper_fmul_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -146,11 +218,19 @@ static bool trans_fdiv_s(DisasContext *ctx,
> arg_fdiv_s *a)
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
>
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fdiv_s(cpu_fpr[a->rd], cpu_env,
> -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
> +    gen_helper_fdiv_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -159,10 +239,16 @@ static bool trans_fsqrt_s(DisasContext *ctx,
> arg_fsqrt_s *a)
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
>
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    check_nanboxed(ctx, 1, t1);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fsqrt_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1]);
> +    gen_helper_fsqrt_s(cpu_fpr[a->rd], cpu_env, t1);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
>      return true;
>  }
>
> @@ -170,14 +256,23 @@ static bool trans_fsgnj_s(DisasContext *ctx,
> arg_fsgnj_s *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
>      if (a->rs1 == a->rs2) { /* FMOV */
> -        tcg_gen_mov_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1]);
> +        tcg_gen_mov_i64(cpu_fpr[a->rd], t1);
>      } else { /* FSGNJ */
> -        tcg_gen_deposit_i64(cpu_fpr[a->rd], cpu_fpr[a->rs2],
> cpu_fpr[a->rs1],
> -                            0, 31);
> +        tcg_gen_deposit_i64(cpu_fpr[a->rd], t2, t1, 0, 31);
>      }
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -185,16 +280,26 @@ static bool trans_fsgnjn_s(DisasContext *ctx,
> arg_fsgnjn_s *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
>      if (a->rs1 == a->rs2) { /* FNEG */
> -        tcg_gen_xori_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], INT32_MIN);
> +        tcg_gen_xori_i64(cpu_fpr[a->rd], t1, INT32_MIN);
>      } else {
>          TCGv_i64 t0 = tcg_temp_new_i64();
> -        tcg_gen_not_i64(t0, cpu_fpr[a->rs2]);
> -        tcg_gen_deposit_i64(cpu_fpr[a->rd], t0, cpu_fpr[a->rs1], 0, 31);
> +        tcg_gen_not_i64(t0, t2);
> +        tcg_gen_deposit_i64(cpu_fpr[a->rd], t0, t1, 0, 31);
>          tcg_temp_free_i64(t0);
>      }
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -202,16 +307,26 @@ static bool trans_fsgnjx_s(DisasContext *ctx,
> arg_fsgnjx_s *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
>      if (a->rs1 == a->rs2) { /* FABS */
> -        tcg_gen_andi_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], ~INT32_MIN);
> +        tcg_gen_andi_i64(cpu_fpr[a->rd], t1, ~INT32_MIN);
>      } else {
>          TCGv_i64 t0 = tcg_temp_new_i64();
> -        tcg_gen_andi_i64(t0, cpu_fpr[a->rs2], INT32_MIN);
> -        tcg_gen_xor_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], t0);
> +        tcg_gen_andi_i64(t0, t2, INT32_MIN);
> +        tcg_gen_xor_i64(cpu_fpr[a->rd], t1, t0);
>          tcg_temp_free_i64(t0);
>      }
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -220,10 +335,18 @@ static bool trans_fmin_s(DisasContext *ctx,
> arg_fmin_s *a)
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
>
> -    gen_helper_fmin_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
> -                      cpu_fpr[a->rs2]);
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
> +    gen_helper_fmin_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -232,10 +355,18 @@ static bool trans_fmax_s(DisasContext *ctx,
> arg_fmax_s *a)
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
>
> -    gen_helper_fmax_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
> -                      cpu_fpr[a->rs2]);
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
> +    gen_helper_fmax_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>      gen_nanbox_fpr(ctx, a->rd);
> +
>      mark_fs_dirty(ctx);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -245,11 +376,16 @@ static bool trans_fcvt_w_s(DisasContext *ctx,
> arg_fcvt_w_s *a)
>      REQUIRE_EXT(ctx, RVF);
>
>      TCGv t0 = tcg_temp_new();
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    check_nanboxed(ctx, 1, t1);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fcvt_w_s(t0, cpu_env, cpu_fpr[a->rs1]);
> +    gen_helper_fcvt_w_s(t0, cpu_env, t1);
>      gen_set_gpr(a->rd, t0);
> -    tcg_temp_free(t0);
>
> +    tcg_temp_free(t0);
> +    tcg_temp_free_i64(t1);
>      return true;
>  }
>
> @@ -259,11 +395,16 @@ static bool trans_fcvt_wu_s(DisasContext *ctx,
> arg_fcvt_wu_s *a)
>      REQUIRE_EXT(ctx, RVF);
>
>      TCGv t0 = tcg_temp_new();
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    check_nanboxed(ctx, 1, t1);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fcvt_wu_s(t0, cpu_env, cpu_fpr[a->rs1]);
> +    gen_helper_fcvt_wu_s(t0, cpu_env, t1);
>      gen_set_gpr(a->rd, t0);
> -    tcg_temp_free(t0);
>
> +    tcg_temp_free(t0);
> +    tcg_temp_free_i64(t1);
>      return true;
>  }
>
> @@ -291,10 +432,20 @@ static bool trans_feq_s(DisasContext *ctx, arg_feq_s
> *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
>      TCGv t0 = tcg_temp_new();
> -    gen_helper_feq_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
> +    gen_helper_feq_s(t0, cpu_env, t1, t2);
>      gen_set_gpr(a->rd, t0);
> +
>      tcg_temp_free(t0);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -302,10 +453,20 @@ static bool trans_flt_s(DisasContext *ctx, arg_flt_s
> *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
>      TCGv t0 = tcg_temp_new();
> -    gen_helper_flt_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
> +    gen_helper_flt_s(t0, cpu_env, t1, t2);
>      gen_set_gpr(a->rd, t0);
> +
>      tcg_temp_free(t0);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -313,10 +474,20 @@ static bool trans_fle_s(DisasContext *ctx, arg_fle_s
> *a)
>  {
>      REQUIRE_FPU;
>      REQUIRE_EXT(ctx, RVF);
> +
>      TCGv t0 = tcg_temp_new();
> -    gen_helper_fle_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    TCGv_i64 t2 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
> +    check_nanboxed(ctx, 2, t1, t2);
> +
> +    gen_helper_fle_s(t0, cpu_env, t1, t2);
>      gen_set_gpr(a->rd, t0);
> +
>      tcg_temp_free(t0);
> +    tcg_temp_free_i64(t1);
> +    tcg_temp_free_i64(t2);
>      return true;
>  }
>
> @@ -326,12 +497,15 @@ static bool trans_fclass_s(DisasContext *ctx,
> arg_fclass_s *a)
>      REQUIRE_EXT(ctx, RVF);
>
>      TCGv t0 = tcg_temp_new();
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    check_nanboxed(ctx, 1, t1);
>
> -    gen_helper_fclass_s(t0, cpu_fpr[a->rs1]);
> -
> +    gen_helper_fclass_s(t0, t1);
>      gen_set_gpr(a->rd, t0);
> -    tcg_temp_free(t0);
>
> +    tcg_temp_free(t0);
> +    tcg_temp_free_i64(t1);
>      return true;
>  }
>
> @@ -400,10 +574,16 @@ static bool trans_fcvt_l_s(DisasContext *ctx,
> arg_fcvt_l_s *a)
>      REQUIRE_EXT(ctx, RVF);
>
>      TCGv t0 = tcg_temp_new();
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    check_nanboxed(ctx, 1, t1);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fcvt_l_s(t0, cpu_env, cpu_fpr[a->rs1]);
> +    gen_helper_fcvt_l_s(t0, cpu_env, t1);
>      gen_set_gpr(a->rd, t0);
> +
>      tcg_temp_free(t0);
> +    tcg_temp_free_i64(t1);
>      return true;
>  }
>
> @@ -413,10 +593,16 @@ static bool trans_fcvt_lu_s(DisasContext *ctx,
> arg_fcvt_lu_s *a)
>      REQUIRE_EXT(ctx, RVF);
>
>      TCGv t0 = tcg_temp_new();
> +    TCGv_i64 t1 = tcg_temp_new_i64();
> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
> +    check_nanboxed(ctx, 1, t1);
> +
>      gen_set_rm(ctx, a->rm);
> -    gen_helper_fcvt_lu_s(t0, cpu_env, cpu_fpr[a->rs1]);
> +    gen_helper_fcvt_lu_s(t0, cpu_env, t1);
>      gen_set_gpr(a->rd, t0);
> +
>      tcg_temp_free(t0);
> +    tcg_temp_free_i64(t1);
>      return true;
>  }
>
> --
> 2.23.0
>
>
It may be more readable to use local macro to wrap allocation and free of
tcg temp variables. Most functions are two-operands,
some requires one and the other needs three operands.  They may be like

#define GEN_ONE_OPERAND \

      TCGv_i64 t1 = tcg_temp_new_i64(); \
      tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
      check_nanboxed(ctx, 1, t1);


 #define GEN_TWO_OPERAND \

      TCGv_i64 t1 = tcg_temp_new_i64(); \

      TCGv_i64 t2 = tcg_temp_new_i64(); \

      tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
      tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]); \
      check_nanboxed(ctx, 2, t1, t2);


  #define GEN_THREE_OPERAND \

      TCGv_i64 t1 = tcg_temp_new_i64(); \

      TCGv_i64 t2 = tcg_temp_new_i64(); \

      TCGv_i64 t3 = tcg_temp_new_i64(); \

      tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
      tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]); \
      tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]); \
      check_nanboxed(ctx, 3, t1, t2, t3);


  #define FREE_ONE_OPERAND \

      tcg_temp_free_i64(t1);



  #define FREE_TWO_OPERAND \

      tcg_temp_free_i64(t1); \

      tcg_temp_free_i64(t2);



  #define FREE_THREE_OPERAND \

      tcg_temp_free_i64(t1); \

      tcg_temp_free_i64(t2); \

      tcg_temp_free_i64(t3);

Chih-Min Chao
LIU Zhiwei June 30, 2020, 7:37 a.m. UTC | #2
On 2020/6/30 15:31, Chih-Min Chao wrote:
> On Sat, Jun 27, 2020 at 5:09 AM LIU Zhiwei <zhiwei_liu@c-sky.com 
> <mailto:zhiwei_liu@c-sky.com>> wrote:
>
>     Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com
>     <mailto:zhiwei_liu@c-sky.com>>
>     ---
>      target/riscv/insn_trans/trans_rvd.inc.c |   7 +-
>      target/riscv/insn_trans/trans_rvf.inc.c | 272
>     ++++++++++++++++++++----
>      2 files changed, 235 insertions(+), 44 deletions(-)
>
>     diff --git a/target/riscv/insn_trans/trans_rvd.inc.c
>     b/target/riscv/insn_trans/trans_rvd.inc.c
>     index c0f4a0c789..16947ea6da 100644
>     --- a/target/riscv/insn_trans/trans_rvd.inc.c
>     +++ b/target/riscv/insn_trans/trans_rvd.inc.c
>     @@ -241,10 +241,15 @@ static bool trans_fcvt_d_s(DisasContext
>     *ctx, arg_fcvt_d_s *a)
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVD);
>
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    check_nanboxed(ctx, 1, t1);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fcvt_d_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1]);
>     +    gen_helper_fcvt_d_s(cpu_fpr[a->rd], cpu_env, t1);
>
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>          return true;
>      }
>
>     diff --git a/target/riscv/insn_trans/trans_rvf.inc.c
>     b/target/riscv/insn_trans/trans_rvf.inc.c
>     index 04bc8e5cb5..b0379b9d1f 100644
>     --- a/target/riscv/insn_trans/trans_rvf.inc.c
>     +++ b/target/riscv/insn_trans/trans_rvf.inc.c
>     @@ -58,11 +58,23 @@ static bool trans_fmadd_s(DisasContext *ctx,
>     arg_fmadd_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    TCGv_i64 t3 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
>     +    check_nanboxed(ctx, 3, t1, t2, t3);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fmadd_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>     -                       cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
>     +    gen_helper_fmadd_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>     +    tcg_temp_free_i64(t3);
>          return true;
>      }
>
>     @@ -70,11 +82,23 @@ static bool trans_fmsub_s(DisasContext *ctx,
>     arg_fmsub_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    TCGv_i64 t3 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
>     +    check_nanboxed(ctx, 3, t1, t2, t3);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fmsub_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>     -                       cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
>     +    gen_helper_fmsub_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>     +    tcg_temp_free_i64(t3);
>          return true;
>      }
>
>     @@ -82,11 +106,23 @@ static bool trans_fnmsub_s(DisasContext *ctx,
>     arg_fnmsub_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    TCGv_i64 t3 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
>     +    check_nanboxed(ctx, 3, t1, t2, t3);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fnmsub_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>     -                        cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
>     +    gen_helper_fnmsub_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>     +    tcg_temp_free_i64(t3);
>          return true;
>      }
>
>     @@ -94,11 +130,23 @@ static bool trans_fnmadd_s(DisasContext *ctx,
>     arg_fnmadd_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    TCGv_i64 t3 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
>     +    check_nanboxed(ctx, 3, t1, t2, t3);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fnmadd_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>     -                        cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
>     -    mark_fs_dirty(ctx);
>     +    gen_helper_fnmadd_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>     +    mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>     +    tcg_temp_free_i64(t3);
>          return true;
>      }
>
>     @@ -107,11 +155,19 @@ static bool trans_fadd_s(DisasContext *ctx,
>     arg_fadd_s *a)
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fadd_s(cpu_fpr[a->rd], cpu_env,
>     -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>     +    gen_helper_fadd_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -120,11 +176,19 @@ static bool trans_fsub_s(DisasContext *ctx,
>     arg_fsub_s *a)
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fsub_s(cpu_fpr[a->rd], cpu_env,
>     -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>     +    gen_helper_fsub_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -133,11 +197,19 @@ static bool trans_fmul_s(DisasContext *ctx,
>     arg_fmul_s *a)
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fmul_s(cpu_fpr[a->rd], cpu_env,
>     -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>     +    gen_helper_fmul_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -146,11 +218,19 @@ static bool trans_fdiv_s(DisasContext *ctx,
>     arg_fdiv_s *a)
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fdiv_s(cpu_fpr[a->rd], cpu_env,
>     -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>     +    gen_helper_fdiv_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -159,10 +239,16 @@ static bool trans_fsqrt_s(DisasContext *ctx,
>     arg_fsqrt_s *a)
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    check_nanboxed(ctx, 1, t1);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fsqrt_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1]);
>     +    gen_helper_fsqrt_s(cpu_fpr[a->rd], cpu_env, t1);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>          return true;
>      }
>
>     @@ -170,14 +256,23 @@ static bool trans_fsgnj_s(DisasContext *ctx,
>     arg_fsgnj_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>          if (a->rs1 == a->rs2) { /* FMOV */
>     -        tcg_gen_mov_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1]);
>     +        tcg_gen_mov_i64(cpu_fpr[a->rd], t1);
>          } else { /* FSGNJ */
>     -        tcg_gen_deposit_i64(cpu_fpr[a->rd], cpu_fpr[a->rs2],
>     cpu_fpr[a->rs1],
>     -                            0, 31);
>     +        tcg_gen_deposit_i64(cpu_fpr[a->rd], t2, t1, 0, 31);
>          }
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -185,16 +280,26 @@ static bool trans_fsgnjn_s(DisasContext
>     *ctx, arg_fsgnjn_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>          if (a->rs1 == a->rs2) { /* FNEG */
>     -        tcg_gen_xori_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], INT32_MIN);
>     +        tcg_gen_xori_i64(cpu_fpr[a->rd], t1, INT32_MIN);
>          } else {
>              TCGv_i64 t0 = tcg_temp_new_i64();
>     -        tcg_gen_not_i64(t0, cpu_fpr[a->rs2]);
>     -        tcg_gen_deposit_i64(cpu_fpr[a->rd], t0, cpu_fpr[a->rs1],
>     0, 31);
>     +        tcg_gen_not_i64(t0, t2);
>     +        tcg_gen_deposit_i64(cpu_fpr[a->rd], t0, t1, 0, 31);
>              tcg_temp_free_i64(t0);
>          }
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -202,16 +307,26 @@ static bool trans_fsgnjx_s(DisasContext
>     *ctx, arg_fsgnjx_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>          if (a->rs1 == a->rs2) { /* FABS */
>     -        tcg_gen_andi_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1],
>     ~INT32_MIN);
>     +        tcg_gen_andi_i64(cpu_fpr[a->rd], t1, ~INT32_MIN);
>          } else {
>              TCGv_i64 t0 = tcg_temp_new_i64();
>     -        tcg_gen_andi_i64(t0, cpu_fpr[a->rs2], INT32_MIN);
>     -        tcg_gen_xor_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], t0);
>     +        tcg_gen_andi_i64(t0, t2, INT32_MIN);
>     +        tcg_gen_xor_i64(cpu_fpr[a->rd], t1, t0);
>              tcg_temp_free_i64(t0);
>          }
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -220,10 +335,18 @@ static bool trans_fmin_s(DisasContext *ctx,
>     arg_fmin_s *a)
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>
>     -    gen_helper_fmin_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>     -                      cpu_fpr[a->rs2]);
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>     +    gen_helper_fmin_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -232,10 +355,18 @@ static bool trans_fmax_s(DisasContext *ctx,
>     arg_fmax_s *a)
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>
>     -    gen_helper_fmax_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>     -                      cpu_fpr[a->rs2]);
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>     +    gen_helper_fmax_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>          gen_nanbox_fpr(ctx, a->rd);
>     +
>          mark_fs_dirty(ctx);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -245,11 +376,16 @@ static bool trans_fcvt_w_s(DisasContext
>     *ctx, arg_fcvt_w_s *a)
>          REQUIRE_EXT(ctx, RVF);
>
>          TCGv t0 = tcg_temp_new();
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    check_nanboxed(ctx, 1, t1);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fcvt_w_s(t0, cpu_env, cpu_fpr[a->rs1]);
>     +    gen_helper_fcvt_w_s(t0, cpu_env, t1);
>          gen_set_gpr(a->rd, t0);
>     -    tcg_temp_free(t0);
>
>     +    tcg_temp_free(t0);
>     +    tcg_temp_free_i64(t1);
>          return true;
>      }
>
>     @@ -259,11 +395,16 @@ static bool trans_fcvt_wu_s(DisasContext
>     *ctx, arg_fcvt_wu_s *a)
>          REQUIRE_EXT(ctx, RVF);
>
>          TCGv t0 = tcg_temp_new();
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    check_nanboxed(ctx, 1, t1);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fcvt_wu_s(t0, cpu_env, cpu_fpr[a->rs1]);
>     +    gen_helper_fcvt_wu_s(t0, cpu_env, t1);
>          gen_set_gpr(a->rd, t0);
>     -    tcg_temp_free(t0);
>
>     +    tcg_temp_free(t0);
>     +    tcg_temp_free_i64(t1);
>          return true;
>      }
>
>     @@ -291,10 +432,20 @@ static bool trans_feq_s(DisasContext *ctx,
>     arg_feq_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>          TCGv t0 = tcg_temp_new();
>     -    gen_helper_feq_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>     +    gen_helper_feq_s(t0, cpu_env, t1, t2);
>          gen_set_gpr(a->rd, t0);
>     +
>          tcg_temp_free(t0);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -302,10 +453,20 @@ static bool trans_flt_s(DisasContext *ctx,
>     arg_flt_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>          TCGv t0 = tcg_temp_new();
>     -    gen_helper_flt_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>     +    gen_helper_flt_s(t0, cpu_env, t1, t2);
>          gen_set_gpr(a->rd, t0);
>     +
>          tcg_temp_free(t0);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -313,10 +474,20 @@ static bool trans_fle_s(DisasContext *ctx,
>     arg_fle_s *a)
>      {
>          REQUIRE_FPU;
>          REQUIRE_EXT(ctx, RVF);
>     +
>          TCGv t0 = tcg_temp_new();
>     -    gen_helper_fle_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    TCGv_i64 t2 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>     +    check_nanboxed(ctx, 2, t1, t2);
>     +
>     +    gen_helper_fle_s(t0, cpu_env, t1, t2);
>          gen_set_gpr(a->rd, t0);
>     +
>          tcg_temp_free(t0);
>     +    tcg_temp_free_i64(t1);
>     +    tcg_temp_free_i64(t2);
>          return true;
>      }
>
>     @@ -326,12 +497,15 @@ static bool trans_fclass_s(DisasContext
>     *ctx, arg_fclass_s *a)
>          REQUIRE_EXT(ctx, RVF);
>
>          TCGv t0 = tcg_temp_new();
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    check_nanboxed(ctx, 1, t1);
>
>     -    gen_helper_fclass_s(t0, cpu_fpr[a->rs1]);
>     -
>     +    gen_helper_fclass_s(t0, t1);
>          gen_set_gpr(a->rd, t0);
>     -    tcg_temp_free(t0);
>
>     +    tcg_temp_free(t0);
>     +    tcg_temp_free_i64(t1);
>          return true;
>      }
>
>     @@ -400,10 +574,16 @@ static bool trans_fcvt_l_s(DisasContext
>     *ctx, arg_fcvt_l_s *a)
>          REQUIRE_EXT(ctx, RVF);
>
>          TCGv t0 = tcg_temp_new();
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    check_nanboxed(ctx, 1, t1);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fcvt_l_s(t0, cpu_env, cpu_fpr[a->rs1]);
>     +    gen_helper_fcvt_l_s(t0, cpu_env, t1);
>          gen_set_gpr(a->rd, t0);
>     +
>          tcg_temp_free(t0);
>     +    tcg_temp_free_i64(t1);
>          return true;
>      }
>
>     @@ -413,10 +593,16 @@ static bool trans_fcvt_lu_s(DisasContext
>     *ctx, arg_fcvt_lu_s *a)
>          REQUIRE_EXT(ctx, RVF);
>
>          TCGv t0 = tcg_temp_new();
>     +    TCGv_i64 t1 = tcg_temp_new_i64();
>     +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>     +    check_nanboxed(ctx, 1, t1);
>     +
>          gen_set_rm(ctx, a->rm);
>     -    gen_helper_fcvt_lu_s(t0, cpu_env, cpu_fpr[a->rs1]);
>     +    gen_helper_fcvt_lu_s(t0, cpu_env, t1);
>          gen_set_gpr(a->rd, t0);
>     +
>          tcg_temp_free(t0);
>     +    tcg_temp_free_i64(t1);
>          return true;
>      }
>
>     -- 
>     2.23.0
>
>
> It may be more readable to use local macro to wrap allocation and free 
> of tcg temp variables. Most functions are two-operands,
> some requires one and the other needs three operands. They may be like
>
> #define GEN_ONE_OPERAND \
>       TCGv_i64 t1 = tcg_temp_new_i64(); \
>       tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
>       check_nanboxed(ctx, 1, t1);
>
>  #define GEN_TWO_OPERAND \
>       TCGv_i64 t1 = tcg_temp_new_i64(); \
>       TCGv_i64 t2 = tcg_temp_new_i64(); \
>       tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
>       tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]); \
>       check_nanboxed(ctx, 2, t1, t2);
>
>   #define GEN_THREE_OPERAND \
>       TCGv_i64 t1 = tcg_temp_new_i64(); \
>       TCGv_i64 t2 = tcg_temp_new_i64(); \
>       TCGv_i64 t3 = tcg_temp_new_i64(); \
>       tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
>       tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]); \
>       tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]); \
>       check_nanboxed(ctx, 3, t1, t2, t3);
>
>   #define FREE_ONE_OPERAND \
>       tcg_temp_free_i64(t1);
>
>   #define FREE_TWO_OPERAND \
>       tcg_temp_free_i64(t1); \
>       tcg_temp_free_i64(t2);
>
>   #define FREE_THREE_OPERAND \
>       tcg_temp_free_i64(t1); \
>       tcg_temp_free_i64(t2); \
>       tcg_temp_free_i64(t3);
>
Good.

Do you think inline function will be better? I just don't like many 
MACROS in one function.

Zhiwei
> Chih-Min Chao
Chih-Min Chao July 2, 2020, 6:29 a.m. UTC | #3
On Tue, Jun 30, 2020 at 3:37 PM LIU Zhiwei <zhiwei_liu@c-sky.com> wrote:

>
>
> On 2020/6/30 15:31, Chih-Min Chao wrote:
>
> On Sat, Jun 27, 2020 at 5:09 AM LIU Zhiwei <zhiwei_liu@c-sky.com> wrote:
>
>> Signed-off-by: LIU Zhiwei <zhiwei_liu@c-sky.com>
>> ---
>>  target/riscv/insn_trans/trans_rvd.inc.c |   7 +-
>>  target/riscv/insn_trans/trans_rvf.inc.c | 272 ++++++++++++++++++++----
>>  2 files changed, 235 insertions(+), 44 deletions(-)
>>
>> diff --git a/target/riscv/insn_trans/trans_rvd.inc.c
>> b/target/riscv/insn_trans/trans_rvd.inc.c
>> index c0f4a0c789..16947ea6da 100644
>> --- a/target/riscv/insn_trans/trans_rvd.inc.c
>> +++ b/target/riscv/insn_trans/trans_rvd.inc.c
>> @@ -241,10 +241,15 @@ static bool trans_fcvt_d_s(DisasContext *ctx,
>> arg_fcvt_d_s *a)
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVD);
>>
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    check_nanboxed(ctx, 1, t1);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fcvt_d_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1]);
>> +    gen_helper_fcvt_d_s(cpu_fpr[a->rd], cpu_env, t1);
>>
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>>      return true;
>>  }
>>
>> diff --git a/target/riscv/insn_trans/trans_rvf.inc.c
>> b/target/riscv/insn_trans/trans_rvf.inc.c
>> index 04bc8e5cb5..b0379b9d1f 100644
>> --- a/target/riscv/insn_trans/trans_rvf.inc.c
>> +++ b/target/riscv/insn_trans/trans_rvf.inc.c
>> @@ -58,11 +58,23 @@ static bool trans_fmadd_s(DisasContext *ctx,
>> arg_fmadd_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    TCGv_i64 t3 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
>> +    check_nanboxed(ctx, 3, t1, t2, t3);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fmadd_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>> -                       cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
>> +    gen_helper_fmadd_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>> +    tcg_temp_free_i64(t3);
>>      return true;
>>  }
>>
>> @@ -70,11 +82,23 @@ static bool trans_fmsub_s(DisasContext *ctx,
>> arg_fmsub_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    TCGv_i64 t3 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
>> +    check_nanboxed(ctx, 3, t1, t2, t3);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fmsub_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>> -                       cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
>> +    gen_helper_fmsub_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>> +    tcg_temp_free_i64(t3);
>>      return true;
>>  }
>>
>> @@ -82,11 +106,23 @@ static bool trans_fnmsub_s(DisasContext *ctx,
>> arg_fnmsub_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    TCGv_i64 t3 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
>> +    check_nanboxed(ctx, 3, t1, t2, t3);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fnmsub_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>> -                        cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
>> +    gen_helper_fnmsub_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>> +    tcg_temp_free_i64(t3);
>>      return true;
>>  }
>>
>> @@ -94,11 +130,23 @@ static bool trans_fnmadd_s(DisasContext *ctx,
>> arg_fnmadd_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    TCGv_i64 t3 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
>> +    check_nanboxed(ctx, 3, t1, t2, t3);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fnmadd_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>> -                        cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
>> -    mark_fs_dirty(ctx);
>> +    gen_helper_fnmadd_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>> +    mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>> +    tcg_temp_free_i64(t3);
>>      return true;
>>  }
>>
>> @@ -107,11 +155,19 @@ static bool trans_fadd_s(DisasContext *ctx,
>> arg_fadd_s *a)
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>>
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fadd_s(cpu_fpr[a->rd], cpu_env,
>> -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>> +    gen_helper_fadd_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -120,11 +176,19 @@ static bool trans_fsub_s(DisasContext *ctx,
>> arg_fsub_s *a)
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>>
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fsub_s(cpu_fpr[a->rd], cpu_env,
>> -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>> +    gen_helper_fsub_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -133,11 +197,19 @@ static bool trans_fmul_s(DisasContext *ctx,
>> arg_fmul_s *a)
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>>
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fmul_s(cpu_fpr[a->rd], cpu_env,
>> -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>> +    gen_helper_fmul_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -146,11 +218,19 @@ static bool trans_fdiv_s(DisasContext *ctx,
>> arg_fdiv_s *a)
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>>
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fdiv_s(cpu_fpr[a->rd], cpu_env,
>> -                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>> +    gen_helper_fdiv_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -159,10 +239,16 @@ static bool trans_fsqrt_s(DisasContext *ctx,
>> arg_fsqrt_s *a)
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>>
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    check_nanboxed(ctx, 1, t1);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fsqrt_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1]);
>> +    gen_helper_fsqrt_s(cpu_fpr[a->rd], cpu_env, t1);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>>      return true;
>>  }
>>
>> @@ -170,14 +256,23 @@ static bool trans_fsgnj_s(DisasContext *ctx,
>> arg_fsgnj_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>>      if (a->rs1 == a->rs2) { /* FMOV */
>> -        tcg_gen_mov_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1]);
>> +        tcg_gen_mov_i64(cpu_fpr[a->rd], t1);
>>      } else { /* FSGNJ */
>> -        tcg_gen_deposit_i64(cpu_fpr[a->rd], cpu_fpr[a->rs2],
>> cpu_fpr[a->rs1],
>> -                            0, 31);
>> +        tcg_gen_deposit_i64(cpu_fpr[a->rd], t2, t1, 0, 31);
>>      }
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -185,16 +280,26 @@ static bool trans_fsgnjn_s(DisasContext *ctx,
>> arg_fsgnjn_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>>      if (a->rs1 == a->rs2) { /* FNEG */
>> -        tcg_gen_xori_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], INT32_MIN);
>> +        tcg_gen_xori_i64(cpu_fpr[a->rd], t1, INT32_MIN);
>>      } else {
>>          TCGv_i64 t0 = tcg_temp_new_i64();
>> -        tcg_gen_not_i64(t0, cpu_fpr[a->rs2]);
>> -        tcg_gen_deposit_i64(cpu_fpr[a->rd], t0, cpu_fpr[a->rs1], 0, 31);
>> +        tcg_gen_not_i64(t0, t2);
>> +        tcg_gen_deposit_i64(cpu_fpr[a->rd], t0, t1, 0, 31);
>
>  t0 is not necessary but t2 could be leveraged directly.

+        tcg_gen_not_i64(t2, t2);
+        tcg_gen_deposit_i64(cpu_fpr[a->rd], t2, t1, 0, 31);

>          tcg_temp_free_i64(t0);
>>      }
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -202,16 +307,26 @@ static bool trans_fsgnjx_s(DisasContext *ctx,
>> arg_fsgnjx_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>>      if (a->rs1 == a->rs2) { /* FABS */
>> -        tcg_gen_andi_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], ~INT32_MIN);
>> +        tcg_gen_andi_i64(cpu_fpr[a->rd], t1, ~INT32_MIN);
>>      } else {
>>          TCGv_i64 t0 = tcg_temp_new_i64();
>> -        tcg_gen_andi_i64(t0, cpu_fpr[a->rs2], INT32_MIN);
>> -        tcg_gen_xor_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], t0);
>> +        tcg_gen_andi_i64(t0, t2, INT32_MIN);
>> +        tcg_gen_xor_i64(cpu_fpr[a->rd], t1, t0);
>>          tcg_temp_free_i64(t0);
>>
> the same as above,  t0 could be removed

>      }
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -220,10 +335,18 @@ static bool trans_fmin_s(DisasContext *ctx,
>> arg_fmin_s *a)
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>>
>> -    gen_helper_fmin_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>> -                      cpu_fpr[a->rs2]);
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>> +    gen_helper_fmin_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -232,10 +355,18 @@ static bool trans_fmax_s(DisasContext *ctx,
>> arg_fmax_s *a)
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>>
>> -    gen_helper_fmax_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
>> -                      cpu_fpr[a->rs2]);
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>> +    gen_helper_fmax_s(cpu_fpr[a->rd], cpu_env, t1, t2);
>>      gen_nanbox_fpr(ctx, a->rd);
>> +
>>      mark_fs_dirty(ctx);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -245,11 +376,16 @@ static bool trans_fcvt_w_s(DisasContext *ctx,
>> arg_fcvt_w_s *a)
>>      REQUIRE_EXT(ctx, RVF);
>>
>>      TCGv t0 = tcg_temp_new();
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    check_nanboxed(ctx, 1, t1);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fcvt_w_s(t0, cpu_env, cpu_fpr[a->rs1]);
>> +    gen_helper_fcvt_w_s(t0, cpu_env, t1);
>>      gen_set_gpr(a->rd, t0);
>> -    tcg_temp_free(t0);
>>
>> +    tcg_temp_free(t0);
>> +    tcg_temp_free_i64(t1);
>>      return true;
>>  }
>>
>> @@ -259,11 +395,16 @@ static bool trans_fcvt_wu_s(DisasContext *ctx,
>> arg_fcvt_wu_s *a)
>>      REQUIRE_EXT(ctx, RVF);
>>
>>      TCGv t0 = tcg_temp_new();
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    check_nanboxed(ctx, 1, t1);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fcvt_wu_s(t0, cpu_env, cpu_fpr[a->rs1]);
>> +    gen_helper_fcvt_wu_s(t0, cpu_env, t1);
>>      gen_set_gpr(a->rd, t0);
>> -    tcg_temp_free(t0);
>>
>> +    tcg_temp_free(t0);
>> +    tcg_temp_free_i64(t1);
>>      return true;
>>  }
>>
>> @@ -291,10 +432,20 @@ static bool trans_feq_s(DisasContext *ctx,
>> arg_feq_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>>      TCGv t0 = tcg_temp_new();
>> -    gen_helper_feq_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>> +    gen_helper_feq_s(t0, cpu_env, t1, t2);
>>      gen_set_gpr(a->rd, t0);
>> +
>>      tcg_temp_free(t0);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -302,10 +453,20 @@ static bool trans_flt_s(DisasContext *ctx,
>> arg_flt_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>>      TCGv t0 = tcg_temp_new();
>> -    gen_helper_flt_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>> +    gen_helper_flt_s(t0, cpu_env, t1, t2);
>>      gen_set_gpr(a->rd, t0);
>> +
>>      tcg_temp_free(t0);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -313,10 +474,20 @@ static bool trans_fle_s(DisasContext *ctx,
>> arg_fle_s *a)
>>  {
>>      REQUIRE_FPU;
>>      REQUIRE_EXT(ctx, RVF);
>> +
>>      TCGv t0 = tcg_temp_new();
>> -    gen_helper_fle_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    TCGv_i64 t2 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
>> +    check_nanboxed(ctx, 2, t1, t2);
>> +
>> +    gen_helper_fle_s(t0, cpu_env, t1, t2);
>>      gen_set_gpr(a->rd, t0);
>> +
>>      tcg_temp_free(t0);
>> +    tcg_temp_free_i64(t1);
>> +    tcg_temp_free_i64(t2);
>>      return true;
>>  }
>>
>> @@ -326,12 +497,15 @@ static bool trans_fclass_s(DisasContext *ctx,
>> arg_fclass_s *a)
>>      REQUIRE_EXT(ctx, RVF);
>>
>>      TCGv t0 = tcg_temp_new();
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    check_nanboxed(ctx, 1, t1);
>>
>> -    gen_helper_fclass_s(t0, cpu_fpr[a->rs1]);
>> -
>> +    gen_helper_fclass_s(t0, t1);
>>      gen_set_gpr(a->rd, t0);
>> -    tcg_temp_free(t0);
>>
>> +    tcg_temp_free(t0);
>> +    tcg_temp_free_i64(t1);
>>      return true;
>>  }
>>
>> @@ -400,10 +574,16 @@ static bool trans_fcvt_l_s(DisasContext *ctx,
>> arg_fcvt_l_s *a)
>>      REQUIRE_EXT(ctx, RVF);
>>
>>      TCGv t0 = tcg_temp_new();
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    check_nanboxed(ctx, 1, t1);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fcvt_l_s(t0, cpu_env, cpu_fpr[a->rs1]);
>> +    gen_helper_fcvt_l_s(t0, cpu_env, t1);
>>      gen_set_gpr(a->rd, t0);
>> +
>>      tcg_temp_free(t0);
>> +    tcg_temp_free_i64(t1);
>>      return true;
>>  }
>>
>> @@ -413,10 +593,16 @@ static bool trans_fcvt_lu_s(DisasContext *ctx,
>> arg_fcvt_lu_s *a)
>>      REQUIRE_EXT(ctx, RVF);
>>
>>      TCGv t0 = tcg_temp_new();
>> +    TCGv_i64 t1 = tcg_temp_new_i64();
>> +    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
>> +    check_nanboxed(ctx, 1, t1);
>> +
>>      gen_set_rm(ctx, a->rm);
>> -    gen_helper_fcvt_lu_s(t0, cpu_env, cpu_fpr[a->rs1]);
>> +    gen_helper_fcvt_lu_s(t0, cpu_env, t1);
>>      gen_set_gpr(a->rd, t0);
>> +
>>      tcg_temp_free(t0);
>> +    tcg_temp_free_i64(t1);
>>      return true;
>>  }
>>
>> --
>> 2.23.0
>>
>>
> It may be more readable to use local macro to wrap allocation and free of
> tcg temp variables. Most functions are two-operands,
> some requires one and the other needs three operands.  They may be like
>
> #define GEN_ONE_OPERAND \
>
>       TCGv_i64 t1 = tcg_temp_new_i64(); \
>       tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
>       check_nanboxed(ctx, 1, t1);
>
>
>  #define GEN_TWO_OPERAND \
>
>       TCGv_i64 t1 = tcg_temp_new_i64(); \
>
>       TCGv_i64 t2 = tcg_temp_new_i64(); \
>
>       tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
>       tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]); \
>       check_nanboxed(ctx, 2, t1, t2);
>
>
>   #define GEN_THREE_OPERAND \
>
>       TCGv_i64 t1 = tcg_temp_new_i64(); \
>
>       TCGv_i64 t2 = tcg_temp_new_i64(); \
>
>       TCGv_i64 t3 = tcg_temp_new_i64(); \
>
>       tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]); \
>       tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]); \
>       tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]); \
>       check_nanboxed(ctx, 3, t1, t2, t3);
>
>
>   #define FREE_ONE_OPERAND \
>
>       tcg_temp_free_i64(t1);
>
>
>
>   #define FREE_TWO_OPERAND \
>
>       tcg_temp_free_i64(t1); \
>
>       tcg_temp_free_i64(t2);
>
>
>
>   #define FREE_THREE_OPERAND \
>
>       tcg_temp_free_i64(t1); \
>
>       tcg_temp_free_i64(t2); \
>
>       tcg_temp_free_i64(t3);
>
> Good.
>
> Do you think inline function will be better? I just don't like many MACROS
> in one function.
>
> Zhiwei
>
> Chih-Min Chao
>
> Either macro or inline is ok to me.  Just want an simple wrapping of
duplicated tcg new and free.

Chih-Min Chao
diff mbox series

Patch

diff --git a/target/riscv/insn_trans/trans_rvd.inc.c b/target/riscv/insn_trans/trans_rvd.inc.c
index c0f4a0c789..16947ea6da 100644
--- a/target/riscv/insn_trans/trans_rvd.inc.c
+++ b/target/riscv/insn_trans/trans_rvd.inc.c
@@ -241,10 +241,15 @@  static bool trans_fcvt_d_s(DisasContext *ctx, arg_fcvt_d_s *a)
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVD);
 
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    check_nanboxed(ctx, 1, t1);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fcvt_d_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1]);
+    gen_helper_fcvt_d_s(cpu_fpr[a->rd], cpu_env, t1);
 
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
     return true;
 }
 
diff --git a/target/riscv/insn_trans/trans_rvf.inc.c b/target/riscv/insn_trans/trans_rvf.inc.c
index 04bc8e5cb5..b0379b9d1f 100644
--- a/target/riscv/insn_trans/trans_rvf.inc.c
+++ b/target/riscv/insn_trans/trans_rvf.inc.c
@@ -58,11 +58,23 @@  static bool trans_fmadd_s(DisasContext *ctx, arg_fmadd_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    TCGv_i64 t3 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
+    check_nanboxed(ctx, 3, t1, t2, t3);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fmadd_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
-                       cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
+    gen_helper_fmadd_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
+    tcg_temp_free_i64(t3);
     return true;
 }
 
@@ -70,11 +82,23 @@  static bool trans_fmsub_s(DisasContext *ctx, arg_fmsub_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    TCGv_i64 t3 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
+    check_nanboxed(ctx, 3, t1, t2, t3);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fmsub_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
-                       cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
+    gen_helper_fmsub_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
+    tcg_temp_free_i64(t3);
     return true;
 }
 
@@ -82,11 +106,23 @@  static bool trans_fnmsub_s(DisasContext *ctx, arg_fnmsub_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    TCGv_i64 t3 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
+    check_nanboxed(ctx, 3, t1, t2, t3);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fnmsub_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
-                        cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
+    gen_helper_fnmsub_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
+    tcg_temp_free_i64(t3);
     return true;
 }
 
@@ -94,11 +130,23 @@  static bool trans_fnmadd_s(DisasContext *ctx, arg_fnmadd_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    TCGv_i64 t3 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    tcg_gen_mov_i64(t3, cpu_fpr[a->rs3]);
+    check_nanboxed(ctx, 3, t1, t2, t3);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fnmadd_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
-                        cpu_fpr[a->rs2], cpu_fpr[a->rs3]);
-    mark_fs_dirty(ctx);
+    gen_helper_fnmadd_s(cpu_fpr[a->rd], cpu_env, t1, t2, t3);
     gen_nanbox_fpr(ctx, a->rd);
+
+    mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
+    tcg_temp_free_i64(t3);
     return true;
 }
 
@@ -107,11 +155,19 @@  static bool trans_fadd_s(DisasContext *ctx, arg_fadd_s *a)
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
 
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fadd_s(cpu_fpr[a->rd], cpu_env,
-                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
+    gen_helper_fadd_s(cpu_fpr[a->rd], cpu_env, t1, t2);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -120,11 +176,19 @@  static bool trans_fsub_s(DisasContext *ctx, arg_fsub_s *a)
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
 
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fsub_s(cpu_fpr[a->rd], cpu_env,
-                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
+    gen_helper_fsub_s(cpu_fpr[a->rd], cpu_env, t1, t2);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -133,11 +197,19 @@  static bool trans_fmul_s(DisasContext *ctx, arg_fmul_s *a)
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
 
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fmul_s(cpu_fpr[a->rd], cpu_env,
-                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
+    gen_helper_fmul_s(cpu_fpr[a->rd], cpu_env, t1, t2);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -146,11 +218,19 @@  static bool trans_fdiv_s(DisasContext *ctx, arg_fdiv_s *a)
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
 
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fdiv_s(cpu_fpr[a->rd], cpu_env,
-                      cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
+    gen_helper_fdiv_s(cpu_fpr[a->rd], cpu_env, t1, t2);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -159,10 +239,16 @@  static bool trans_fsqrt_s(DisasContext *ctx, arg_fsqrt_s *a)
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
 
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    check_nanboxed(ctx, 1, t1);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fsqrt_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1]);
+    gen_helper_fsqrt_s(cpu_fpr[a->rd], cpu_env, t1);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
     return true;
 }
 
@@ -170,14 +256,23 @@  static bool trans_fsgnj_s(DisasContext *ctx, arg_fsgnj_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
     if (a->rs1 == a->rs2) { /* FMOV */
-        tcg_gen_mov_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1]);
+        tcg_gen_mov_i64(cpu_fpr[a->rd], t1);
     } else { /* FSGNJ */
-        tcg_gen_deposit_i64(cpu_fpr[a->rd], cpu_fpr[a->rs2], cpu_fpr[a->rs1],
-                            0, 31);
+        tcg_gen_deposit_i64(cpu_fpr[a->rd], t2, t1, 0, 31);
     }
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -185,16 +280,26 @@  static bool trans_fsgnjn_s(DisasContext *ctx, arg_fsgnjn_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
     if (a->rs1 == a->rs2) { /* FNEG */
-        tcg_gen_xori_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], INT32_MIN);
+        tcg_gen_xori_i64(cpu_fpr[a->rd], t1, INT32_MIN);
     } else {
         TCGv_i64 t0 = tcg_temp_new_i64();
-        tcg_gen_not_i64(t0, cpu_fpr[a->rs2]);
-        tcg_gen_deposit_i64(cpu_fpr[a->rd], t0, cpu_fpr[a->rs1], 0, 31);
+        tcg_gen_not_i64(t0, t2);
+        tcg_gen_deposit_i64(cpu_fpr[a->rd], t0, t1, 0, 31);
         tcg_temp_free_i64(t0);
     }
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -202,16 +307,26 @@  static bool trans_fsgnjx_s(DisasContext *ctx, arg_fsgnjx_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
     if (a->rs1 == a->rs2) { /* FABS */
-        tcg_gen_andi_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], ~INT32_MIN);
+        tcg_gen_andi_i64(cpu_fpr[a->rd], t1, ~INT32_MIN);
     } else {
         TCGv_i64 t0 = tcg_temp_new_i64();
-        tcg_gen_andi_i64(t0, cpu_fpr[a->rs2], INT32_MIN);
-        tcg_gen_xor_i64(cpu_fpr[a->rd], cpu_fpr[a->rs1], t0);
+        tcg_gen_andi_i64(t0, t2, INT32_MIN);
+        tcg_gen_xor_i64(cpu_fpr[a->rd], t1, t0);
         tcg_temp_free_i64(t0);
     }
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -220,10 +335,18 @@  static bool trans_fmin_s(DisasContext *ctx, arg_fmin_s *a)
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
 
-    gen_helper_fmin_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
-                      cpu_fpr[a->rs2]);
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
+    gen_helper_fmin_s(cpu_fpr[a->rd], cpu_env, t1, t2);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -232,10 +355,18 @@  static bool trans_fmax_s(DisasContext *ctx, arg_fmax_s *a)
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
 
-    gen_helper_fmax_s(cpu_fpr[a->rd], cpu_env, cpu_fpr[a->rs1],
-                      cpu_fpr[a->rs2]);
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
+    gen_helper_fmax_s(cpu_fpr[a->rd], cpu_env, t1, t2);
     gen_nanbox_fpr(ctx, a->rd);
+
     mark_fs_dirty(ctx);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -245,11 +376,16 @@  static bool trans_fcvt_w_s(DisasContext *ctx, arg_fcvt_w_s *a)
     REQUIRE_EXT(ctx, RVF);
 
     TCGv t0 = tcg_temp_new();
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    check_nanboxed(ctx, 1, t1);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fcvt_w_s(t0, cpu_env, cpu_fpr[a->rs1]);
+    gen_helper_fcvt_w_s(t0, cpu_env, t1);
     gen_set_gpr(a->rd, t0);
-    tcg_temp_free(t0);
 
+    tcg_temp_free(t0);
+    tcg_temp_free_i64(t1);
     return true;
 }
 
@@ -259,11 +395,16 @@  static bool trans_fcvt_wu_s(DisasContext *ctx, arg_fcvt_wu_s *a)
     REQUIRE_EXT(ctx, RVF);
 
     TCGv t0 = tcg_temp_new();
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    check_nanboxed(ctx, 1, t1);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fcvt_wu_s(t0, cpu_env, cpu_fpr[a->rs1]);
+    gen_helper_fcvt_wu_s(t0, cpu_env, t1);
     gen_set_gpr(a->rd, t0);
-    tcg_temp_free(t0);
 
+    tcg_temp_free(t0);
+    tcg_temp_free_i64(t1);
     return true;
 }
 
@@ -291,10 +432,20 @@  static bool trans_feq_s(DisasContext *ctx, arg_feq_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
     TCGv t0 = tcg_temp_new();
-    gen_helper_feq_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
+    gen_helper_feq_s(t0, cpu_env, t1, t2);
     gen_set_gpr(a->rd, t0);
+
     tcg_temp_free(t0);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -302,10 +453,20 @@  static bool trans_flt_s(DisasContext *ctx, arg_flt_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
     TCGv t0 = tcg_temp_new();
-    gen_helper_flt_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
+    gen_helper_flt_s(t0, cpu_env, t1, t2);
     gen_set_gpr(a->rd, t0);
+
     tcg_temp_free(t0);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -313,10 +474,20 @@  static bool trans_fle_s(DisasContext *ctx, arg_fle_s *a)
 {
     REQUIRE_FPU;
     REQUIRE_EXT(ctx, RVF);
+
     TCGv t0 = tcg_temp_new();
-    gen_helper_fle_s(t0, cpu_env, cpu_fpr[a->rs1], cpu_fpr[a->rs2]);
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    TCGv_i64 t2 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    tcg_gen_mov_i64(t2, cpu_fpr[a->rs2]);
+    check_nanboxed(ctx, 2, t1, t2);
+
+    gen_helper_fle_s(t0, cpu_env, t1, t2);
     gen_set_gpr(a->rd, t0);
+
     tcg_temp_free(t0);
+    tcg_temp_free_i64(t1);
+    tcg_temp_free_i64(t2);
     return true;
 }
 
@@ -326,12 +497,15 @@  static bool trans_fclass_s(DisasContext *ctx, arg_fclass_s *a)
     REQUIRE_EXT(ctx, RVF);
 
     TCGv t0 = tcg_temp_new();
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    check_nanboxed(ctx, 1, t1);
 
-    gen_helper_fclass_s(t0, cpu_fpr[a->rs1]);
-
+    gen_helper_fclass_s(t0, t1);
     gen_set_gpr(a->rd, t0);
-    tcg_temp_free(t0);
 
+    tcg_temp_free(t0);
+    tcg_temp_free_i64(t1);
     return true;
 }
 
@@ -400,10 +574,16 @@  static bool trans_fcvt_l_s(DisasContext *ctx, arg_fcvt_l_s *a)
     REQUIRE_EXT(ctx, RVF);
 
     TCGv t0 = tcg_temp_new();
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    check_nanboxed(ctx, 1, t1);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fcvt_l_s(t0, cpu_env, cpu_fpr[a->rs1]);
+    gen_helper_fcvt_l_s(t0, cpu_env, t1);
     gen_set_gpr(a->rd, t0);
+
     tcg_temp_free(t0);
+    tcg_temp_free_i64(t1);
     return true;
 }
 
@@ -413,10 +593,16 @@  static bool trans_fcvt_lu_s(DisasContext *ctx, arg_fcvt_lu_s *a)
     REQUIRE_EXT(ctx, RVF);
 
     TCGv t0 = tcg_temp_new();
+    TCGv_i64 t1 = tcg_temp_new_i64();
+    tcg_gen_mov_i64(t1, cpu_fpr[a->rs1]);
+    check_nanboxed(ctx, 1, t1);
+
     gen_set_rm(ctx, a->rm);
-    gen_helper_fcvt_lu_s(t0, cpu_env, cpu_fpr[a->rs1]);
+    gen_helper_fcvt_lu_s(t0, cpu_env, t1);
     gen_set_gpr(a->rd, t0);
+
     tcg_temp_free(t0);
+    tcg_temp_free_i64(t1);
     return true;
 }