/dragonfly/contrib/gcc-4.7/gcc/ |
H A D | value-prof.c | 563 tree tmp0, tmp1, tmp2, tmpv; in gimple_divmod_fixed_value() local 582 tmp0 = make_ssa_name (tmpv, NULL); in gimple_divmod_fixed_value() 585 SSA_NAME_DEF_STMT (tmp0) = stmt1; in gimple_divmod_fixed_value() 596 op1, tmp0); in gimple_divmod_fixed_value() 1147 tree tmp0, tmp1, tmpv, tmp; in gimple_ic() local 1158 tmp0 = make_ssa_name (tmpv, NULL); in gimple_ic() 1161 load_stmt = gimple_build_assign (tmp0, tmp); in gimple_ic() 1162 SSA_NAME_DEF_STMT (tmp0) = load_stmt; in gimple_ic() 1402 tree tmp0, tmp1, tmpv, vcall_size, optype; in gimple_stringop_fixed_value() local 1420 tmp0 = make_ssa_name (tmpv, NULL); in gimple_stringop_fixed_value() [all …]
|
H A D | loop-iv.c | 2256 rtx op0, op1, delta, step, bound, may_xform, tmp, tmp0, tmp1; in iv_number_of_iterations() local 2567 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode); in iv_number_of_iterations() 2570 SImode, mode, tmp0, tmp1); in iv_number_of_iterations() 2631 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode); in iv_number_of_iterations() 2647 t0 = simplify_gen_binary (UMOD, mode, copy_rtx (tmp0), step); in iv_number_of_iterations() 2665 SImode, mode, tmp0, tmp); in iv_number_of_iterations() 2668 delta = simplify_gen_binary (MINUS, mode, delta, tmp0); in iv_number_of_iterations() 2676 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode); in iv_number_of_iterations() 2689 bound, tmp0); in iv_number_of_iterations() 2701 bound, tmp0); in iv_number_of_iterations() [all …]
|
H A D | simplify-rtx.c | 3433 long tmp0[4]; in simplify_const_binary_operation() local 3438 real_to_target (tmp0, CONST_DOUBLE_REAL_VALUE (op0), in simplify_const_binary_operation() 3447 tmp0[i] &= tmp1[i]; in simplify_const_binary_operation() 3450 tmp0[i] |= tmp1[i]; in simplify_const_binary_operation() 3453 tmp0[i] ^= tmp1[i]; in simplify_const_binary_operation() 3459 real_from_target (&r, tmp0, mode); in simplify_const_binary_operation()
|
H A D | fold-const.c | 10375 tree tmp0 = var0; in fold_binary_loc() local 10378 if (TREE_CODE (tmp0) == NEGATE_EXPR) in fold_binary_loc() 10379 tmp0 = TREE_OPERAND (tmp0, 0); in fold_binary_loc() 10384 if (!operand_equal_p (tmp0, tmp1, 0)) in fold_binary_loc() 10390 tree tmp0 = fold_convert (type, lit0); in fold_binary_loc() local 10393 if (!TREE_OVERFLOW (tmp0) && !TREE_OVERFLOW (tmp1) in fold_binary_loc() 10394 && TREE_OVERFLOW (fold_build2 (code, type, tmp0, tmp1))) in fold_binary_loc()
|
/dragonfly/contrib/gcc-8.0/gcc/ |
H A D | vtable-verify.c | 671 tree tmp0; in verify_bb_vtables() local 760 tmp0 = make_temp_ssa_name (TREE_TYPE (lhs), NULL, "VTV"); in verify_bb_vtables() 761 gimple_call_set_lhs (call_stmt, tmp0); in verify_bb_vtables() 774 SET_USE (use_p, tmp0); in verify_bb_vtables()
|
H A D | value-prof.c | 691 tree tmp0, tmp1, tmp2; in gimple_divmod_fixed_value() local 709 tmp0 = make_temp_ssa_name (optype, NULL, "PROF"); in gimple_divmod_fixed_value() 711 stmt1 = gimple_build_assign (tmp0, fold_convert (optype, value)); in gimple_divmod_fixed_value() 713 stmt3 = gimple_build_cond (NE_EXPR, tmp1, tmp0, NULL_TREE, NULL_TREE); in gimple_divmod_fixed_value() 1303 tree tmp0, tmp1, tmp; in gimple_ic() local 1319 tmp0 = make_temp_ssa_name (optype, NULL, "PROF"); in gimple_ic() 1322 load_stmt = gimple_build_assign (tmp0, tmp); in gimple_ic() 1329 cond_stmt = gimple_build_cond (EQ_EXPR, tmp1, tmp0, NULL_TREE, NULL_TREE); in gimple_ic() 1625 tree tmp0, tmp1, vcall_size, optype; in gimple_stringop_fixed_value() local 1640 tmp0 = make_temp_ssa_name (optype, NULL, "PROF"); in gimple_stringop_fixed_value() [all …]
|
H A D | loop-iv.c | 2309 rtx op0, op1, delta, step, bound, may_xform, tmp, tmp0, tmp1; in iv_number_of_iterations() local 2619 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode); in iv_number_of_iterations() 2622 SImode, mode, tmp0, tmp1); in iv_number_of_iterations() 2683 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode); in iv_number_of_iterations() 2699 t0 = simplify_gen_binary (UMOD, mode, copy_rtx (tmp0), step); in iv_number_of_iterations() 2717 SImode, mode, tmp0, tmp); in iv_number_of_iterations() 2720 delta = simplify_gen_binary (MINUS, mode, delta, tmp0); in iv_number_of_iterations() 2728 tmp0 = lowpart_subreg (mode, iv0.base, comp_mode); in iv_number_of_iterations() 2741 bound, tmp0); in iv_number_of_iterations() 2753 bound, tmp0); in iv_number_of_iterations() [all …]
|
H A D | simplify-rtx.c | 4113 long tmp0[4]; in simplify_const_binary_operation() local 4118 real_to_target (tmp0, CONST_DOUBLE_REAL_VALUE (op0), in simplify_const_binary_operation() 4127 tmp0[i] &= tmp1[i]; in simplify_const_binary_operation() 4130 tmp0[i] |= tmp1[i]; in simplify_const_binary_operation() 4133 tmp0[i] ^= tmp1[i]; in simplify_const_binary_operation() 4139 real_from_target (&r, tmp0, mode); in simplify_const_binary_operation()
|
H A D | fold-const.c | 9749 tree tmp0 = var0 ? var0 : minus_var0; in fold_binary_loc() local 9753 if (TREE_CODE (tmp0) == NEGATE_EXPR) in fold_binary_loc() 9755 tmp0 = TREE_OPERAND (tmp0, 0); in fold_binary_loc() 9758 if (CONVERT_EXPR_P (tmp0) in fold_binary_loc() 9759 && INTEGRAL_TYPE_P (TREE_TYPE (TREE_OPERAND (tmp0, 0))) in fold_binary_loc() 9760 && (TYPE_PRECISION (TREE_TYPE (TREE_OPERAND (tmp0, 0))) in fold_binary_loc() 9762 tmp0 = TREE_OPERAND (tmp0, 0); in fold_binary_loc() 9776 || !operand_equal_p (tmp0, tmp1, 0)) in fold_binary_loc()
|
/dragonfly/sys/crypto/curve25519/ |
H A D | curve25519.c | 855 fe tmp0, tmp1; in curve25519() local 886 fe_sq_tl(&tmp0, &tmp1l); in curve25519() 890 fe_mul_ttt(&x2, &tmp1, &tmp0); in curve25519() 891 fe_sub(&tmp1l, &tmp1, &tmp0); in curve25519() 895 fe_add(&tmp0l, &tmp0, &z3); in curve25519()
|
/dragonfly/crypto/libressl/crypto/curve25519/ |
H A D | curve25519.c | 4815 fe x1, x2, z2, x3, z3, tmp0, tmp1; in x25519_scalar_mult_generic() local 4836 fe_sub(tmp0, x3, z3); in x25519_scalar_mult_generic() 4840 fe_mul(z3, tmp0, x2); in x25519_scalar_mult_generic() 4842 fe_sq(tmp0, tmp1); in x25519_scalar_mult_generic() 4846 fe_mul(x2, tmp1, tmp0); in x25519_scalar_mult_generic() 4847 fe_sub(tmp1, tmp1, tmp0); in x25519_scalar_mult_generic() 4851 fe_add(tmp0, tmp0, z3); in x25519_scalar_mult_generic() 4853 fe_mul(z2, tmp1, tmp0); in x25519_scalar_mult_generic()
|
/dragonfly/crypto/libressl/crypto/ec/ |
H A D | ecp_smpl.c | 1233 BIGNUM *tmp0, *tmp1; in ec_GFp_simple_points_make_affine() local 1248 if ((tmp0 = BN_CTX_get(ctx)) == NULL) in ec_GFp_simple_points_make_affine() 1340 if (!group->meth->field_mul(group, tmp0, heap[i / 2], heap[i + 1], ctx)) in ec_GFp_simple_points_make_affine() 1344 if (!BN_copy(heap[i], tmp0)) in ec_GFp_simple_points_make_affine()
|
/dragonfly/contrib/gcc-4.7/gcc/config/i386/ |
H A D | sse.md | 3049 rtx tmp0, tmp1; 3053 tmp0 = gen_reg_rtx (V4DFmode); 3061 tmp0 = gen_reg_rtx (V4SFmode); 3094 rtx tmp0, tmp1; 3098 tmp0 = gen_reg_rtx (V4DFmode); 3106 tmp0 = gen_reg_rtx (V4SImode); 3175 rtx tmp0, tmp1; 3179 tmp0 = gen_reg_rtx (V4DFmode); 3187 tmp0 = gen_reg_rtx (V4SImode); 9920 rtx tmp0, tmp1; [all …]
|
H A D | i386.c | 16444 rtx scratch, tmp0, tmp1, tmp2; in ix86_split_idivmod() local 16476 tmp0 = gen_rtx_REG (CCNOmode, FLAGS_REG); in ix86_split_idivmod() 16477 tmp0 = gen_rtx_EQ (VOIDmode, tmp0, const0_rtx); in ix86_split_idivmod() 16478 tmp0 = gen_rtx_IF_THEN_ELSE (VOIDmode, tmp0, in ix86_split_idivmod() 16481 insn = emit_jump_insn (gen_rtx_SET (VOIDmode, pc_rtx, tmp0)); in ix86_split_idivmod() 16498 tmp0 = simplify_gen_subreg (HImode, scratch, mode, 0); in ix86_split_idivmod() 16501 emit_insn (gen_udivmodhiqi3 (tmp0, tmp1, tmp2)); in ix86_split_idivmod() 16529 tmp1 = gen_lowpart (QImode, tmp0); in ix86_split_idivmod() 34727 rtx tmp0 = gen_reg_rtx (XFmode); in ix86_emit_i387_round() local 34729 emit_insn (gen_frndintxf2_floor (tmp0, tmp1)); in ix86_emit_i387_round() [all …]
|
/dragonfly/sys/dev/drm/amd/display/dc/dml/ |
H A D | dml1_display_rq_dlg_calc.c | 142 double tmp0 = (max_num_sw * swath_height) / (l_sw - (prefill - 3.0) / 2.0); in get_vratio_pre() local 144 if (tmp0 > vratio_pre) in get_vratio_pre() 145 vratio_pre = tmp0; in get_vratio_pre()
|
/dragonfly/contrib/gcc-8.0/gcc/config/i386/ |
H A D | i386.c | 20761 rtx scratch, tmp0, tmp1, tmp2; in ix86_split_idivmod() local 20806 tmp0 = gen_rtx_REG (CCNOmode, FLAGS_REG); in ix86_split_idivmod() 20807 tmp0 = gen_rtx_EQ (VOIDmode, tmp0, const0_rtx); in ix86_split_idivmod() 20808 tmp0 = gen_rtx_IF_THEN_ELSE (VOIDmode, tmp0, in ix86_split_idivmod() 20828 tmp0 = lowpart_subreg (HImode, scratch, mode); in ix86_split_idivmod() 20853 tmp0, GEN_INT (8), GEN_INT (8)); in ix86_split_idivmod() 20867 tmp1 = gen_lowpart (QImode, tmp0); in ix86_split_idivmod() 45017 rtx tmp0 = gen_reg_rtx (XFmode); in ix86_emit_i387_round() local 45019 emit_insn (gen_frndintxf2_floor (tmp0, tmp1)); in ix86_emit_i387_round() 45022 gen_rtx_UNSPEC (outmode, gen_rtvec (1, tmp0), in ix86_emit_i387_round() [all …]
|