/* * RISC-V translation routines for the RV64D Standard Extension. * * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu * Copyright (c) 2018 Peer Adelt, peer.adelt@hni.uni-paderborn.de * Bastian Koppelmann, kbastian@mail.uni-paderborn.de * * This program is free software; you can redistribute it and/or modify it * under the terms and conditions of the GNU General Public License, * version 2 or later, as published by the Free Software Foundation. * * This program is distributed in the hope it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for * more details. * * You should have received a copy of the GNU General Public License along with * this program. If not, see . */ static bool trans_fld(DisasContext *ctx, arg_fld *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; TCGv t0 = tcg_temp_new(tcg_ctx); gen_get_gpr(ctx, t0, a->rs1); REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); tcg_gen_addi_tl(tcg_ctx, t0, t0, a->imm); tcg_gen_qemu_ld_i64(ctx->uc, tcg_ctx->cpu_fpr_risc[a->rd], t0, ctx->mem_idx, MO_TEQ); mark_fs_dirty(ctx); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fsd(DisasContext *ctx, arg_fsd *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; TCGv t0 = tcg_temp_new(tcg_ctx); gen_get_gpr(ctx, t0, a->rs1); REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); tcg_gen_addi_tl(tcg_ctx, t0, t0, a->imm); tcg_gen_qemu_st_i64(ctx->uc, tcg_ctx->cpu_fpr_risc[a->rs2], t0, ctx->mem_idx, MO_TEQ); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fmadd_d(DisasContext *ctx, arg_fmadd_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fmadd_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2], tcg_ctx->cpu_fpr_risc[a->rs3]); mark_fs_dirty(ctx); return true; } static bool trans_fmsub_d(DisasContext *ctx, arg_fmsub_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fmsub_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2], tcg_ctx->cpu_fpr_risc[a->rs3]); mark_fs_dirty(ctx); return true; } static bool trans_fnmsub_d(DisasContext *ctx, arg_fnmsub_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fnmsub_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2], tcg_ctx->cpu_fpr_risc[a->rs3]); mark_fs_dirty(ctx); return true; } static bool trans_fnmadd_d(DisasContext *ctx, arg_fnmadd_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fnmadd_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2], tcg_ctx->cpu_fpr_risc[a->rs3]); mark_fs_dirty(ctx); return true; } static bool trans_fadd_d(DisasContext *ctx, arg_fadd_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fadd_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); mark_fs_dirty(ctx); return true; } static bool trans_fsub_d(DisasContext *ctx, arg_fsub_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fsub_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); mark_fs_dirty(ctx); return true; } static bool trans_fmul_d(DisasContext *ctx, arg_fmul_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fmul_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); mark_fs_dirty(ctx); return true; } static bool trans_fdiv_d(DisasContext *ctx, arg_fdiv_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fdiv_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); mark_fs_dirty(ctx); return true; } static bool trans_fsqrt_d(DisasContext *ctx, arg_fsqrt_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fsqrt_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1]); mark_fs_dirty(ctx); return true; } static bool trans_fsgnj_d(DisasContext *ctx, arg_fsgnj_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; if (a->rs1 == a->rs2) { /* FMOV */ tcg_gen_mov_i64(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_fpr_risc[a->rs1]); } else { tcg_gen_deposit_i64(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_fpr_risc[a->rs2], tcg_ctx->cpu_fpr_risc[a->rs1], 0, 63); } mark_fs_dirty(ctx); return true; } static bool trans_fsgnjn_d(DisasContext *ctx, arg_fsgnjn_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); if (a->rs1 == a->rs2) { /* FNEG */ tcg_gen_xori_i64(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_fpr_risc[a->rs1], INT64_MIN); } else { TCGv_i64 t0 = tcg_temp_new_i64(tcg_ctx); tcg_gen_not_i64(tcg_ctx, t0, tcg_ctx->cpu_fpr_risc[a->rs2]); tcg_gen_deposit_i64(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], t0, tcg_ctx->cpu_fpr_risc[a->rs1], 0, 63); tcg_temp_free_i64(tcg_ctx, t0); } mark_fs_dirty(ctx); return true; } static bool trans_fsgnjx_d(DisasContext *ctx, arg_fsgnjx_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); if (a->rs1 == a->rs2) { /* FABS */ tcg_gen_andi_i64(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_fpr_risc[a->rs1], ~INT64_MIN); } else { TCGv_i64 t0 = tcg_temp_new_i64(tcg_ctx); tcg_gen_andi_i64(tcg_ctx, t0, tcg_ctx->cpu_fpr_risc[a->rs2], INT64_MIN); tcg_gen_xor_i64(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_fpr_risc[a->rs1], t0); tcg_temp_free_i64(tcg_ctx, t0); } mark_fs_dirty(ctx); return true; } static bool trans_fmin_d(DisasContext *ctx, arg_fmin_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_helper_fmin_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); mark_fs_dirty(ctx); return true; } static bool trans_fmax_d(DisasContext *ctx, arg_fmax_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_helper_fmax_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); mark_fs_dirty(ctx); return true; } static bool trans_fcvt_s_d(DisasContext *ctx, arg_fcvt_s_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fcvt_s_d(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1]); mark_fs_dirty(ctx); return true; } static bool trans_fcvt_d_s(DisasContext *ctx, arg_fcvt_d_s *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_rm(ctx, a->rm); gen_helper_fcvt_d_s(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1]); mark_fs_dirty(ctx); return true; } static bool trans_feq_d(DisasContext *ctx, arg_feq_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_helper_feq_d(tcg_ctx, t0, tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); gen_set_gpr(ctx, a->rd, t0); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_flt_d(DisasContext *ctx, arg_flt_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_helper_flt_d(tcg_ctx, t0, tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); gen_set_gpr(ctx, a->rd, t0); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fle_d(DisasContext *ctx, arg_fle_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_helper_fle_d(tcg_ctx, t0, tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1], tcg_ctx->cpu_fpr_risc[a->rs2]); gen_set_gpr(ctx, a->rd, t0); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fclass_d(DisasContext *ctx, arg_fclass_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_helper_fclass_d(tcg_ctx, t0, tcg_ctx->cpu_fpr_risc[a->rs1]); gen_set_gpr(ctx, a->rd, t0); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fcvt_w_d(DisasContext *ctx, arg_fcvt_w_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_set_rm(ctx, a->rm); gen_helper_fcvt_w_d(tcg_ctx, t0, tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1]); gen_set_gpr(ctx, a->rd, t0); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fcvt_wu_d(DisasContext *ctx, arg_fcvt_wu_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_set_rm(ctx, a->rm); gen_helper_fcvt_wu_d(tcg_ctx, t0, tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1]); gen_set_gpr(ctx, a->rd, t0); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fcvt_d_w(DisasContext *ctx, arg_fcvt_d_w *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_get_gpr(ctx, t0, a->rs1); gen_set_rm(ctx, a->rm); gen_helper_fcvt_d_w(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, t0); tcg_temp_free(tcg_ctx, t0); mark_fs_dirty(ctx); return true; } static bool trans_fcvt_d_wu(DisasContext *ctx, arg_fcvt_d_wu *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_get_gpr(ctx, t0, a->rs1); gen_set_rm(ctx, a->rm); gen_helper_fcvt_d_wu(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, t0); tcg_temp_free(tcg_ctx, t0); mark_fs_dirty(ctx); return true; } #ifdef TARGET_RISCV64 static bool trans_fcvt_l_d(DisasContext *ctx, arg_fcvt_l_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_set_rm(ctx, a->rm); gen_helper_fcvt_l_d(tcg_ctx, t0, tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1]); gen_set_gpr(ctx, a->rd, t0); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fcvt_lu_d(DisasContext *ctx, arg_fcvt_lu_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_set_rm(ctx, a->rm); gen_helper_fcvt_lu_d(tcg_ctx, t0, tcg_ctx->cpu_env, tcg_ctx->cpu_fpr_risc[a->rs1]); gen_set_gpr(ctx, a->rd, t0); tcg_temp_free(tcg_ctx, t0); return true; } static bool trans_fmv_x_d(DisasContext *ctx, arg_fmv_x_d *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); gen_set_gpr(ctx, a->rd, tcg_ctx->cpu_fpr_risc[a->rs1]); return true; } static bool trans_fcvt_d_l(DisasContext *ctx, arg_fcvt_d_l *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_get_gpr(ctx, t0, a->rs1); gen_set_rm(ctx, a->rm); gen_helper_fcvt_d_l(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, t0); tcg_temp_free(tcg_ctx, t0); mark_fs_dirty(ctx); return true; } static bool trans_fcvt_d_lu(DisasContext *ctx, arg_fcvt_d_lu *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_get_gpr(ctx, t0, a->rs1); gen_set_rm(ctx, a->rm); gen_helper_fcvt_d_lu(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], tcg_ctx->cpu_env, t0); tcg_temp_free(tcg_ctx, t0); mark_fs_dirty(ctx); return true; } static bool trans_fmv_d_x(DisasContext *ctx, arg_fmv_d_x *a) { TCGContext *tcg_ctx = ctx->uc->tcg_ctx; REQUIRE_FPU; REQUIRE_EXT(ctx, RVD); TCGv t0 = tcg_temp_new(tcg_ctx); gen_get_gpr(ctx, t0, a->rs1); tcg_gen_mov_tl(tcg_ctx, tcg_ctx->cpu_fpr_risc[a->rd], t0); tcg_temp_free(tcg_ctx, t0); mark_fs_dirty(ctx); return true; } #endif