diff options
author | matz <matz@138bc75d-0d04-0410-961f-82ee72b054a4> | 2009-08-27 12:42:18 +0000 |
---|---|---|
committer | matz <matz@138bc75d-0d04-0410-961f-82ee72b054a4> | 2009-08-27 12:42:18 +0000 |
commit | dff12ad7a582d6cd2141e858481a25d85e212886 (patch) | |
tree | 5a9300b4b8ddc05b1473f2097f2165eadb64a37b /gcc/expr.c | |
parent | c6f21ec7fd18880cc4f879d0ffddf5565a44bfea (diff) | |
download | gcc-dff12ad7a582d6cd2141e858481a25d85e212886.tar.gz |
* expr.c (expand_expr_real_2): New function taking exploded
unary or binary expression, split out from ...
(expand_expr_real_1): ... here. Move over all unary/binary
switch parts to above function, in particular these codes:
PAREN_EXPR, NOP_EXPR, CONVERT_EXPR, POINTER_PLUS_EXPR, PLUS_EXPR,
MINUS_EXPR, MULT_EXPR, TRUNC_DIV_EXPR, FLOOR_DIV_EXPR, CEIL_DIV_EXPR,
ROUND_DIV_EXPR, EXACT_DIV_EXPR, RDIV_EXPR, TRUNC_MOD_EXPR,
FLOOR_MOD_EXPR, CEIL_MOD_EXPR, ROUND_MOD_EXPR, FIXED_CONVERT_EXPR,
FIX_TRUNC_EXPR, FLOAT_EXPR, NEGATE_EXPR, ABS_EXPR, MAX_EXPR, MIN_EXPR,
BIT_NOT_EXPR, TRUTH_AND_EXPR, BIT_AND_EXPR, TRUTH_OR_EXPR,
BIT_IOR_EXPR, TRUTH_XOR_EXPR, BIT_XOR_EXPR, LROTATE_EXPR, RROTATE_EXPR,
LSHIFT_EXPR, RSHIFT_EXPR, LT_EXPR, LE_EXPR, GT_EXPR, GE_EXPR, EQ_EXPR,
NE_EXPR, UNORDERED_EXPR, ORDERED_EXPR, UNLT_EXPR, UNLE_EXPR, UNGT_EXPR,
UNGE_EXPR, UNEQ_EXPR, LTGT_EXPR, TRUTH_NOT_EXPR, COMPLEX_EXPR,
WIDEN_SUM_EXPR, REDUC_MAX_EXPR, REDUC_MIN_EXPR, REDUC_PLUS_EXPR,
VEC_EXTRACT_EVEN_EXPR, VEC_EXTRACT_ODD_EXPR, VEC_INTERLEAVE_HIGH_EXPR,
VEC_INTERLEAVE_LOW_EXPR, VEC_LSHIFT_EXPR, VEC_RSHIFT_EXPR,
VEC_UNPACK_HI_EXPR, VEC_UNPACK_LO_EXPR, VEC_UNPACK_FLOAT_HI_EXPR,
VEC_UNPACK_FLOAT_LO_EXPR, VEC_WIDEN_MULT_HI_EXPR,
VEC_WIDEN_MULT_LO_EXPR, VEC_PACK_TRUNC_EXPR, VEC_PACK_SAT_EXPR,
VEC_PACK_FIX_TRUNC_EXPR.
(<case PAREN_EXPR>): Call set_mem_attributes() with type, not the
full expression.
git-svn-id: svn+ssh://gcc.gnu.org/svn/gcc/trunk@151142 138bc75d-0d04-0410-961f-82ee72b054a4
Diffstat (limited to 'gcc/expr.c')
-rw-r--r-- | gcc/expr.c | 2016 |
1 files changed, 1049 insertions, 967 deletions
diff --git a/gcc/expr.c b/gcc/expr.c index a753682a150..a906506f07b 100644 --- a/gcc/expr.c +++ b/gcc/expr.c @@ -7204,23 +7204,22 @@ expand_expr_real (tree exp, rtx target, enum machine_mode tmode, } static rtx -expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, - enum expand_modifier modifier, rtx *alt_rtl) +expand_expr_real_2 (sepops ops, rtx target, enum machine_mode tmode, + enum expand_modifier modifier) { - rtx op0, op1, op2, temp, decl_rtl; + rtx op0, op1, op2, temp; tree type; int unsignedp; enum machine_mode mode; - enum tree_code code = TREE_CODE (exp); + enum tree_code code = ops->code; optab this_optab; rtx subtarget, original_target; int ignore; - tree context, subexp0, subexp1; + tree subexp0, subexp1; bool reduce_bit_field; gimple subexp0_def, subexp1_def; tree top0, top1; - location_t loc = EXPR_LOCATION (exp); - struct separate_ops ops; + location_t loc = ops->location; tree treeop0, treeop1, treeop2; #define REDUCE_BIT_FIELD(expr) (reduce_bit_field \ ? reduce_to_bit_field_precision ((expr), \ @@ -7228,6 +7227,1048 @@ expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, type) \ : (expr)) + type = ops->type; + mode = TYPE_MODE (type); + unsignedp = TYPE_UNSIGNED (type); + + treeop0 = ops->op0; + treeop1 = ops->op1; + treeop2 = ops->op2; + + /* We should be called only on simple (binary or unary) expressions, + exactly those that are valid in gimple expressions that aren't + GIMPLE_SINGLE_RHS (or invalid). */ + gcc_assert (get_gimple_rhs_class (code) == GIMPLE_UNARY_RHS + || get_gimple_rhs_class (code) == GIMPLE_BINARY_RHS); + + ignore = (target == const0_rtx + || ((CONVERT_EXPR_CODE_P (code) + || code == COND_EXPR || code == VIEW_CONVERT_EXPR) + && TREE_CODE (type) == VOID_TYPE)); + + /* We should be called only if we need the result. */ + gcc_assert (!ignore); + + /* An operation in what may be a bit-field type needs the + result to be reduced to the precision of the bit-field type, + which is narrower than that of the type's mode. */ + reduce_bit_field = (TREE_CODE (type) == INTEGER_TYPE + && GET_MODE_PRECISION (mode) > TYPE_PRECISION (type)); + + if (reduce_bit_field && modifier == EXPAND_STACK_PARM) + target = 0; + + /* Use subtarget as the target for operand 0 of a binary operation. */ + subtarget = get_subtarget (target); + original_target = target; + + switch (code) + { + case PAREN_EXPR: + CASE_CONVERT: + if (treeop0 == error_mark_node) + return const0_rtx; + + if (TREE_CODE (type) == UNION_TYPE) + { + tree valtype = TREE_TYPE (treeop0); + + /* If both input and output are BLKmode, this conversion isn't doing + anything except possibly changing memory attribute. */ + if (mode == BLKmode && TYPE_MODE (valtype) == BLKmode) + { + rtx result = expand_expr (treeop0, target, tmode, + modifier); + + result = copy_rtx (result); + set_mem_attributes (result, type, 0); + return result; + } + + if (target == 0) + { + if (TYPE_MODE (type) != BLKmode) + target = gen_reg_rtx (TYPE_MODE (type)); + else + target = assign_temp (type, 0, 1, 1); + } + + if (MEM_P (target)) + /* Store data into beginning of memory target. */ + store_expr (treeop0, + adjust_address (target, TYPE_MODE (valtype), 0), + modifier == EXPAND_STACK_PARM, + false); + + else + { + gcc_assert (REG_P (target)); + + /* Store this field into a union of the proper type. */ + store_field (target, + MIN ((int_size_in_bytes (TREE_TYPE + (treeop0)) + * BITS_PER_UNIT), + (HOST_WIDE_INT) GET_MODE_BITSIZE (mode)), + 0, TYPE_MODE (valtype), treeop0, + type, 0, false); + } + + /* Return the entire union. */ + return target; + } + + if (mode == TYPE_MODE (TREE_TYPE (treeop0))) + { + op0 = expand_expr (treeop0, target, VOIDmode, + modifier); + + /* If the signedness of the conversion differs and OP0 is + a promoted SUBREG, clear that indication since we now + have to do the proper extension. */ + if (TYPE_UNSIGNED (TREE_TYPE (treeop0)) != unsignedp + && GET_CODE (op0) == SUBREG) + SUBREG_PROMOTED_VAR_P (op0) = 0; + + return REDUCE_BIT_FIELD (op0); + } + + op0 = expand_expr (treeop0, NULL_RTX, mode, + modifier == EXPAND_SUM ? EXPAND_NORMAL : modifier); + if (GET_MODE (op0) == mode) + ; + + /* If OP0 is a constant, just convert it into the proper mode. */ + else if (CONSTANT_P (op0)) + { + tree inner_type = TREE_TYPE (treeop0); + enum machine_mode inner_mode = TYPE_MODE (inner_type); + + if (modifier == EXPAND_INITIALIZER) + op0 = simplify_gen_subreg (mode, op0, inner_mode, + subreg_lowpart_offset (mode, + inner_mode)); + else + op0= convert_modes (mode, inner_mode, op0, + TYPE_UNSIGNED (inner_type)); + } + + else if (modifier == EXPAND_INITIALIZER) + op0 = gen_rtx_fmt_e (unsignedp ? ZERO_EXTEND : SIGN_EXTEND, mode, op0); + + else if (target == 0) + op0 = convert_to_mode (mode, op0, + TYPE_UNSIGNED (TREE_TYPE + (treeop0))); + else + { + convert_move (target, op0, + TYPE_UNSIGNED (TREE_TYPE (treeop0))); + op0 = target; + } + + return REDUCE_BIT_FIELD (op0); + + case POINTER_PLUS_EXPR: + /* Even though the sizetype mode and the pointer's mode can be different + expand is able to handle this correctly and get the correct result out + of the PLUS_EXPR code. */ + /* Make sure to sign-extend the sizetype offset in a POINTER_PLUS_EXPR + if sizetype precision is smaller than pointer precision. */ + if (TYPE_PRECISION (sizetype) < TYPE_PRECISION (type)) + treeop1 = fold_convert_loc (loc, type, + fold_convert_loc (loc, ssizetype, + treeop1)); + case PLUS_EXPR: + + /* Check if this is a case for multiplication and addition. */ + if ((TREE_CODE (type) == INTEGER_TYPE + || TREE_CODE (type) == FIXED_POINT_TYPE) + && (subexp0_def = get_def_for_expr (treeop0, + MULT_EXPR))) + { + tree subsubexp0, subsubexp1; + gimple subsubexp0_def, subsubexp1_def; + enum tree_code this_code; + + this_code = TREE_CODE (type) == INTEGER_TYPE ? NOP_EXPR + : FIXED_CONVERT_EXPR; + subsubexp0 = gimple_assign_rhs1 (subexp0_def); + subsubexp0_def = get_def_for_expr (subsubexp0, this_code); + subsubexp1 = gimple_assign_rhs2 (subexp0_def); + subsubexp1_def = get_def_for_expr (subsubexp1, this_code); + if (subsubexp0_def && subsubexp1_def + && (top0 = gimple_assign_rhs1 (subsubexp0_def)) + && (top1 = gimple_assign_rhs1 (subsubexp1_def)) + && (TYPE_PRECISION (TREE_TYPE (top0)) + < TYPE_PRECISION (TREE_TYPE (subsubexp0))) + && (TYPE_PRECISION (TREE_TYPE (top0)) + == TYPE_PRECISION (TREE_TYPE (top1))) + && (TYPE_UNSIGNED (TREE_TYPE (top0)) + == TYPE_UNSIGNED (TREE_TYPE (top1)))) + { + tree op0type = TREE_TYPE (top0); + enum machine_mode innermode = TYPE_MODE (op0type); + bool zextend_p = TYPE_UNSIGNED (op0type); + bool sat_p = TYPE_SATURATING (TREE_TYPE (subsubexp0)); + if (sat_p == 0) + this_optab = zextend_p ? umadd_widen_optab : smadd_widen_optab; + else + this_optab = zextend_p ? usmadd_widen_optab + : ssmadd_widen_optab; + if (mode == GET_MODE_2XWIDER_MODE (innermode) + && (optab_handler (this_optab, mode)->insn_code + != CODE_FOR_nothing)) + { + expand_operands (top0, top1, NULL_RTX, &op0, &op1, + EXPAND_NORMAL); + op2 = expand_expr (treeop1, subtarget, + VOIDmode, EXPAND_NORMAL); + temp = expand_ternary_op (mode, this_optab, op0, op1, op2, + target, unsignedp); + gcc_assert (temp); + return REDUCE_BIT_FIELD (temp); + } + } + } + + /* If we are adding a constant, a VAR_DECL that is sp, fp, or ap, and + something else, make sure we add the register to the constant and + then to the other thing. This case can occur during strength + reduction and doing it this way will produce better code if the + frame pointer or argument pointer is eliminated. + + fold-const.c will ensure that the constant is always in the inner + PLUS_EXPR, so the only case we need to do anything about is if + sp, ap, or fp is our second argument, in which case we must swap + the innermost first argument and our second argument. */ + + if (TREE_CODE (treeop0) == PLUS_EXPR + && TREE_CODE (TREE_OPERAND (treeop0, 1)) == INTEGER_CST + && TREE_CODE (treeop1) == VAR_DECL + && (DECL_RTL (treeop1) == frame_pointer_rtx + || DECL_RTL (treeop1) == stack_pointer_rtx + || DECL_RTL (treeop1) == arg_pointer_rtx)) + { + tree t = treeop1; + + treeop1 = TREE_OPERAND (treeop0, 0); + TREE_OPERAND (treeop0, 0) = t; + } + + /* If the result is to be ptr_mode and we are adding an integer to + something, we might be forming a constant. So try to use + plus_constant. If it produces a sum and we can't accept it, + use force_operand. This allows P = &ARR[const] to generate + efficient code on machines where a SYMBOL_REF is not a valid + address. + + If this is an EXPAND_SUM call, always return the sum. */ + if (modifier == EXPAND_SUM || modifier == EXPAND_INITIALIZER + || (mode == ptr_mode && (unsignedp || ! flag_trapv))) + { + if (modifier == EXPAND_STACK_PARM) + target = 0; + if (TREE_CODE (treeop0) == INTEGER_CST + && GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_WIDE_INT + && TREE_CONSTANT (treeop1)) + { + rtx constant_part; + + op1 = expand_expr (treeop1, subtarget, VOIDmode, + EXPAND_SUM); + /* Use immed_double_const to ensure that the constant is + truncated according to the mode of OP1, then sign extended + to a HOST_WIDE_INT. Using the constant directly can result + in non-canonical RTL in a 64x32 cross compile. */ + constant_part + = immed_double_const (TREE_INT_CST_LOW (treeop0), + (HOST_WIDE_INT) 0, + TYPE_MODE (TREE_TYPE (treeop1))); + op1 = plus_constant (op1, INTVAL (constant_part)); + if (modifier != EXPAND_SUM && modifier != EXPAND_INITIALIZER) + op1 = force_operand (op1, target); + return REDUCE_BIT_FIELD (op1); + } + + else if (TREE_CODE (treeop1) == INTEGER_CST + && GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_WIDE_INT + && TREE_CONSTANT (treeop0)) + { + rtx constant_part; + + op0 = expand_expr (treeop0, subtarget, VOIDmode, + (modifier == EXPAND_INITIALIZER + ? EXPAND_INITIALIZER : EXPAND_SUM)); + if (! CONSTANT_P (op0)) + { + op1 = expand_expr (treeop1, NULL_RTX, + VOIDmode, modifier); + /* Return a PLUS if modifier says it's OK. */ + if (modifier == EXPAND_SUM + || modifier == EXPAND_INITIALIZER) + return simplify_gen_binary (PLUS, mode, op0, op1); + goto binop2; + } + /* Use immed_double_const to ensure that the constant is + truncated according to the mode of OP1, then sign extended + to a HOST_WIDE_INT. Using the constant directly can result + in non-canonical RTL in a 64x32 cross compile. */ + constant_part + = immed_double_const (TREE_INT_CST_LOW (treeop1), + (HOST_WIDE_INT) 0, + TYPE_MODE (TREE_TYPE (treeop0))); + op0 = plus_constant (op0, INTVAL (constant_part)); + if (modifier != EXPAND_SUM && modifier != EXPAND_INITIALIZER) + op0 = force_operand (op0, target); + return REDUCE_BIT_FIELD (op0); + } + } + + /* No sense saving up arithmetic to be done + if it's all in the wrong mode to form part of an address. + And force_operand won't know whether to sign-extend or + zero-extend. */ + if ((modifier != EXPAND_SUM && modifier != EXPAND_INITIALIZER) + || mode != ptr_mode) + { + expand_operands (treeop0, treeop1, + subtarget, &op0, &op1, EXPAND_NORMAL); + if (op0 == const0_rtx) + return op1; + if (op1 == const0_rtx) + return op0; + goto binop2; + } + + expand_operands (treeop0, treeop1, + subtarget, &op0, &op1, modifier); + return REDUCE_BIT_FIELD (simplify_gen_binary (PLUS, mode, op0, op1)); + + case MINUS_EXPR: + /* Check if this is a case for multiplication and subtraction. */ + if ((TREE_CODE (type) == INTEGER_TYPE + || TREE_CODE (type) == FIXED_POINT_TYPE) + && (subexp1_def = get_def_for_expr (treeop1, + MULT_EXPR))) + { + tree subsubexp0, subsubexp1; + gimple subsubexp0_def, subsubexp1_def; + enum tree_code this_code; + + this_code = TREE_CODE (type) == INTEGER_TYPE ? NOP_EXPR + : FIXED_CONVERT_EXPR; + subsubexp0 = gimple_assign_rhs1 (subexp1_def); + subsubexp0_def = get_def_for_expr (subsubexp0, this_code); + subsubexp1 = gimple_assign_rhs2 (subexp1_def); + subsubexp1_def = get_def_for_expr (subsubexp1, this_code); + if (subsubexp0_def && subsubexp1_def + && (top0 = gimple_assign_rhs1 (subsubexp0_def)) + && (top1 = gimple_assign_rhs1 (subsubexp1_def)) + && (TYPE_PRECISION (TREE_TYPE (top0)) + < TYPE_PRECISION (TREE_TYPE (subsubexp0))) + && (TYPE_PRECISION (TREE_TYPE (top0)) + == TYPE_PRECISION (TREE_TYPE (top1))) + && (TYPE_UNSIGNED (TREE_TYPE (top0)) + == TYPE_UNSIGNED (TREE_TYPE (top1)))) + { + tree op0type = TREE_TYPE (top0); + enum machine_mode innermode = TYPE_MODE (op0type); + bool zextend_p = TYPE_UNSIGNED (op0type); + bool sat_p = TYPE_SATURATING (TREE_TYPE (subsubexp0)); + if (sat_p == 0) + this_optab = zextend_p ? umsub_widen_optab : smsub_widen_optab; + else + this_optab = zextend_p ? usmsub_widen_optab + : ssmsub_widen_optab; + if (mode == GET_MODE_2XWIDER_MODE (innermode) + && (optab_handler (this_optab, mode)->insn_code + != CODE_FOR_nothing)) + { + expand_operands (top0, top1, NULL_RTX, &op0, &op1, + EXPAND_NORMAL); + op2 = expand_expr (treeop0, subtarget, + VOIDmode, EXPAND_NORMAL); + temp = expand_ternary_op (mode, this_optab, op0, op1, op2, + target, unsignedp); + gcc_assert (temp); + return REDUCE_BIT_FIELD (temp); + } + } + } + + /* For initializers, we are allowed to return a MINUS of two + symbolic constants. Here we handle all cases when both operands + are constant. */ + /* Handle difference of two symbolic constants, + for the sake of an initializer. */ + if ((modifier == EXPAND_SUM || modifier == EXPAND_INITIALIZER) + && really_constant_p (treeop0) + && really_constant_p (treeop1)) + { + expand_operands (treeop0, treeop1, + NULL_RTX, &op0, &op1, modifier); + + /* If the last operand is a CONST_INT, use plus_constant of + the negated constant. Else make the MINUS. */ + if (CONST_INT_P (op1)) + return REDUCE_BIT_FIELD (plus_constant (op0, - INTVAL (op1))); + else + return REDUCE_BIT_FIELD (gen_rtx_MINUS (mode, op0, op1)); + } + + /* No sense saving up arithmetic to be done + if it's all in the wrong mode to form part of an address. + And force_operand won't know whether to sign-extend or + zero-extend. */ + if ((modifier != EXPAND_SUM && modifier != EXPAND_INITIALIZER) + || mode != ptr_mode) + goto binop; + + expand_operands (treeop0, treeop1, + subtarget, &op0, &op1, modifier); + + /* Convert A - const to A + (-const). */ + if (CONST_INT_P (op1)) + { + op1 = negate_rtx (mode, op1); + return REDUCE_BIT_FIELD (simplify_gen_binary (PLUS, mode, op0, op1)); + } + + goto binop2; + + case MULT_EXPR: + /* If this is a fixed-point operation, then we cannot use the code + below because "expand_mult" doesn't support sat/no-sat fixed-point + multiplications. */ + if (ALL_FIXED_POINT_MODE_P (mode)) + goto binop; + + /* If first operand is constant, swap them. + Thus the following special case checks need only + check the second operand. */ + if (TREE_CODE (treeop0) == INTEGER_CST) + { + tree t1 = treeop0; + treeop0 = treeop1; + treeop1 = t1; + } + + /* Attempt to return something suitable for generating an + indexed address, for machines that support that. */ + + if (modifier == EXPAND_SUM && mode == ptr_mode + && host_integerp (treeop1, 0)) + { + tree exp1 = treeop1; + + op0 = expand_expr (treeop0, subtarget, VOIDmode, + EXPAND_SUM); + + if (!REG_P (op0)) + op0 = force_operand (op0, NULL_RTX); + if (!REG_P (op0)) + op0 = copy_to_mode_reg (mode, op0); + + return REDUCE_BIT_FIELD (gen_rtx_MULT (mode, op0, + gen_int_mode (tree_low_cst (exp1, 0), + TYPE_MODE (TREE_TYPE (exp1))))); + } + + if (modifier == EXPAND_STACK_PARM) + target = 0; + + /* Check for multiplying things that have been extended + from a narrower type. If this machine supports multiplying + in that narrower type with a result in the desired type, + do it that way, and avoid the explicit type-conversion. */ + + subexp0 = treeop0; + subexp1 = treeop1; + subexp0_def = get_def_for_expr (subexp0, NOP_EXPR); + subexp1_def = get_def_for_expr (subexp1, NOP_EXPR); + top0 = top1 = NULL_TREE; + + /* First, check if we have a multiplication of one signed and one + unsigned operand. */ + if (subexp0_def + && (top0 = gimple_assign_rhs1 (subexp0_def)) + && subexp1_def + && (top1 = gimple_assign_rhs1 (subexp1_def)) + && TREE_CODE (type) == INTEGER_TYPE + && (TYPE_PRECISION (TREE_TYPE (top0)) + < TYPE_PRECISION (TREE_TYPE (subexp0))) + && (TYPE_PRECISION (TREE_TYPE (top0)) + == TYPE_PRECISION (TREE_TYPE (top1))) + && (TYPE_UNSIGNED (TREE_TYPE (top0)) + != TYPE_UNSIGNED (TREE_TYPE (top1)))) + { + enum machine_mode innermode + = TYPE_MODE (TREE_TYPE (top0)); + this_optab = usmul_widen_optab; + if (mode == GET_MODE_WIDER_MODE (innermode)) + { + if (optab_handler (this_optab, mode)->insn_code != CODE_FOR_nothing) + { + if (TYPE_UNSIGNED (TREE_TYPE (top0))) + expand_operands (top0, top1, NULL_RTX, &op0, &op1, + EXPAND_NORMAL); + else + expand_operands (top0, top1, NULL_RTX, &op1, &op0, + EXPAND_NORMAL); + + goto binop3; + } + } + } + /* Check for a multiplication with matching signedness. If + valid, TOP0 and TOP1 were set in the previous if + condition. */ + else if (top0 + && TREE_CODE (type) == INTEGER_TYPE + && (TYPE_PRECISION (TREE_TYPE (top0)) + < TYPE_PRECISION (TREE_TYPE (subexp0))) + && ((TREE_CODE (subexp1) == INTEGER_CST + && int_fits_type_p (subexp1, TREE_TYPE (top0)) + /* Don't use a widening multiply if a shift will do. */ + && ((GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (subexp1))) + > HOST_BITS_PER_WIDE_INT) + || exact_log2 (TREE_INT_CST_LOW (subexp1)) < 0)) + || + (top1 + && (TYPE_PRECISION (TREE_TYPE (top1)) + == TYPE_PRECISION (TREE_TYPE (top0)) + /* If both operands are extended, they must either both + be zero-extended or both be sign-extended. */ + && (TYPE_UNSIGNED (TREE_TYPE (top1)) + == TYPE_UNSIGNED (TREE_TYPE (top0))))))) + { + tree op0type = TREE_TYPE (top0); + enum machine_mode innermode = TYPE_MODE (op0type); + bool zextend_p = TYPE_UNSIGNED (op0type); + optab other_optab = zextend_p ? smul_widen_optab : umul_widen_optab; + this_optab = zextend_p ? umul_widen_optab : smul_widen_optab; + + if (mode == GET_MODE_2XWIDER_MODE (innermode)) + { + if (optab_handler (this_optab, mode)->insn_code != CODE_FOR_nothing) + { + if (TREE_CODE (subexp1) == INTEGER_CST) + expand_operands (top0, subexp1, NULL_RTX, &op0, &op1, + EXPAND_NORMAL); + else + expand_operands (top0, top1, NULL_RTX, &op0, &op1, + EXPAND_NORMAL); + goto binop3; + } + else if (optab_handler (other_optab, mode)->insn_code != CODE_FOR_nothing + && innermode == word_mode) + { + rtx htem, hipart; + op0 = expand_normal (top0); + if (TREE_CODE (subexp1) == INTEGER_CST) + op1 = convert_modes (innermode, mode, + expand_normal (subexp1), unsignedp); + else + op1 = expand_normal (top1); + temp = expand_binop (mode, other_optab, op0, op1, target, + unsignedp, OPTAB_LIB_WIDEN); + hipart = gen_highpart (innermode, temp); + htem = expand_mult_highpart_adjust (innermode, hipart, + op0, op1, hipart, + zextend_p); + if (htem != hipart) + emit_move_insn (hipart, htem); + return REDUCE_BIT_FIELD (temp); + } + } + } + expand_operands (subexp0, subexp1, subtarget, &op0, &op1, EXPAND_NORMAL); + return REDUCE_BIT_FIELD (expand_mult (mode, op0, op1, target, unsignedp)); + + case TRUNC_DIV_EXPR: + case FLOOR_DIV_EXPR: + case CEIL_DIV_EXPR: + case ROUND_DIV_EXPR: + case EXACT_DIV_EXPR: + /* If this is a fixed-point operation, then we cannot use the code + below because "expand_divmod" doesn't support sat/no-sat fixed-point + divisions. */ + if (ALL_FIXED_POINT_MODE_P (mode)) + goto binop; + + if (modifier == EXPAND_STACK_PARM) + target = 0; + /* Possible optimization: compute the dividend with EXPAND_SUM + then if the divisor is constant can optimize the case + where some terms of the dividend have coeffs divisible by it. */ + expand_operands (treeop0, treeop1, + subtarget, &op0, &op1, EXPAND_NORMAL); + return expand_divmod (0, code, mode, op0, op1, target, unsignedp); + + case RDIV_EXPR: + goto binop; + + case TRUNC_MOD_EXPR: + case FLOOR_MOD_EXPR: + case CEIL_MOD_EXPR: + case ROUND_MOD_EXPR: + if (modifier == EXPAND_STACK_PARM) + target = 0; + expand_operands (treeop0, treeop1, + subtarget, &op0, &op1, EXPAND_NORMAL); + return expand_divmod (1, code, mode, op0, op1, target, unsignedp); + + case FIXED_CONVERT_EXPR: + op0 = expand_normal (treeop0); + if (target == 0 || modifier == EXPAND_STACK_PARM) + target = gen_reg_rtx (mode); + + if ((TREE_CODE (TREE_TYPE (treeop0)) == INTEGER_TYPE + && TYPE_UNSIGNED (TREE_TYPE (treeop0))) + || (TREE_CODE (type) == INTEGER_TYPE && TYPE_UNSIGNED (type))) + expand_fixed_convert (target, op0, 1, TYPE_SATURATING (type)); + else + expand_fixed_convert (target, op0, 0, TYPE_SATURATING (type)); + return target; + + case FIX_TRUNC_EXPR: + op0 = expand_normal (treeop0); + if (target == 0 || modifier == EXPAND_STACK_PARM) + target = gen_reg_rtx (mode); + expand_fix (target, op0, unsignedp); + return target; + + case FLOAT_EXPR: + op0 = expand_normal (treeop0); + if (target == 0 || modifier == EXPAND_STACK_PARM) + target = gen_reg_rtx (mode); + /* expand_float can't figure out what to do if FROM has VOIDmode. + So give it the correct mode. With -O, cse will optimize this. */ + if (GET_MODE (op0) == VOIDmode) + op0 = copy_to_mode_reg (TYPE_MODE (TREE_TYPE (treeop0)), + op0); + expand_float (target, op0, + TYPE_UNSIGNED (TREE_TYPE (treeop0))); + return target; + + case NEGATE_EXPR: + op0 = expand_expr (treeop0, subtarget, + VOIDmode, EXPAND_NORMAL); + if (modifier == EXPAND_STACK_PARM) + target = 0; + temp = expand_unop (mode, + optab_for_tree_code (NEGATE_EXPR, type, + optab_default), + op0, target, 0); + gcc_assert (temp); + return REDUCE_BIT_FIELD (temp); + + case ABS_EXPR: + op0 = expand_expr (treeop0, subtarget, + VOIDmode, EXPAND_NORMAL); + if (modifier == EXPAND_STACK_PARM) + target = 0; + + /* ABS_EXPR is not valid for complex arguments. */ + gcc_assert (GET_MODE_CLASS (mode) != MODE_COMPLEX_INT + && GET_MODE_CLASS (mode) != MODE_COMPLEX_FLOAT); + + /* Unsigned abs is simply the operand. Testing here means we don't + risk generating incorrect code below. */ + if (TYPE_UNSIGNED (type)) + return op0; + + return expand_abs (mode, op0, target, unsignedp, + safe_from_p (target, treeop0, 1)); + + case MAX_EXPR: + case MIN_EXPR: + target = original_target; + if (target == 0 + || modifier == EXPAND_STACK_PARM + || (MEM_P (target) && MEM_VOLATILE_P (target)) + || GET_MODE (target) != mode + || (REG_P (target) + && REGNO (target) < FIRST_PSEUDO_REGISTER)) + target = gen_reg_rtx (mode); + expand_operands (treeop0, treeop1, + target, &op0, &op1, EXPAND_NORMAL); + + /* First try to do it with a special MIN or MAX instruction. + If that does not win, use a conditional jump to select the proper + value. */ + this_optab = optab_for_tree_code (code, type, optab_default); + temp = expand_binop (mode, this_optab, op0, op1, target, unsignedp, + OPTAB_WIDEN); + if (temp != 0) + return temp; + + /* At this point, a MEM target is no longer useful; we will get better + code without it. */ + + if (! REG_P (target)) + target = gen_reg_rtx (mode); + + /* If op1 was placed in target, swap op0 and op1. */ + if (target != op0 && target == op1) + { + temp = op0; + op0 = op1; + op1 = temp; + } + + /* We generate better code and avoid problems with op1 mentioning + target by forcing op1 into a pseudo if it isn't a constant. */ + if (! CONSTANT_P (op1)) + op1 = force_reg (mode, op1); + + { + enum rtx_code comparison_code; + rtx cmpop1 = op1; + + if (code == MAX_EXPR) + comparison_code = unsignedp ? GEU : GE; + else + comparison_code = unsignedp ? LEU : LE; + + /* Canonicalize to comparisons against 0. */ + if (op1 == const1_rtx) + { + /* Converting (a >= 1 ? a : 1) into (a > 0 ? a : 1) + or (a != 0 ? a : 1) for unsigned. + For MIN we are safe converting (a <= 1 ? a : 1) + into (a <= 0 ? a : 1) */ + cmpop1 = const0_rtx; + if (code == MAX_EXPR) + comparison_code = unsignedp ? NE : GT; + } + if (op1 == constm1_rtx && !unsignedp) + { + /* Converting (a >= -1 ? a : -1) into (a >= 0 ? a : -1) + and (a <= -1 ? a : -1) into (a < 0 ? a : -1) */ + cmpop1 = const0_rtx; + if (code == MIN_EXPR) + comparison_code = LT; + } +#ifdef HAVE_conditional_move + /* Use a conditional move if possible. */ + if (can_conditionally_move_p (mode)) + { + rtx insn; + + /* ??? Same problem as in expmed.c: emit_conditional_move + forces a stack adjustment via compare_from_rtx, and we + lose the stack adjustment if the sequence we are about + to create is discarded. */ + do_pending_stack_adjust (); + + start_sequence (); + + /* Try to emit the conditional move. */ + insn = emit_conditional_move (target, comparison_code, + op0, cmpop1, mode, + op0, op1, mode, + unsignedp); + + /* If we could do the conditional move, emit the sequence, + and return. */ + if (insn) + { + rtx seq = get_insns (); + end_sequence (); + emit_insn (seq); + return target; + } + + /* Otherwise discard the sequence and fall back to code with + branches. */ + end_sequence (); + } +#endif + if (target != op0) + emit_move_insn (target, op0); + + temp = gen_label_rtx (); + do_compare_rtx_and_jump (target, cmpop1, comparison_code, + unsignedp, mode, NULL_RTX, NULL_RTX, temp); + } + emit_move_insn (target, op1); + emit_label (temp); + return target; + + case BIT_NOT_EXPR: + op0 = expand_expr (treeop0, subtarget, + VOIDmode, EXPAND_NORMAL); + if (modifier == EXPAND_STACK_PARM) + target = 0; + temp = expand_unop (mode, one_cmpl_optab, op0, target, 1); + gcc_assert (temp); + return temp; + + /* ??? Can optimize bitwise operations with one arg constant. + Can optimize (a bitwise1 n) bitwise2 (a bitwise3 b) + and (a bitwise1 b) bitwise2 b (etc) + but that is probably not worth while. */ + + /* BIT_AND_EXPR is for bitwise anding. TRUTH_AND_EXPR is for anding two + boolean values when we want in all cases to compute both of them. In + general it is fastest to do TRUTH_AND_EXPR by computing both operands + as actual zero-or-1 values and then bitwise anding. In cases where + there cannot be any side effects, better code would be made by + treating TRUTH_AND_EXPR like TRUTH_ANDIF_EXPR; but the question is + how to recognize those cases. */ + + case TRUTH_AND_EXPR: + code = BIT_AND_EXPR; + case BIT_AND_EXPR: + goto binop; + + case TRUTH_OR_EXPR: + code = BIT_IOR_EXPR; + case BIT_IOR_EXPR: + goto binop; + + case TRUTH_XOR_EXPR: + code = BIT_XOR_EXPR; + case BIT_XOR_EXPR: + goto binop; + + case LROTATE_EXPR: + case RROTATE_EXPR: + gcc_assert (VECTOR_MODE_P (TYPE_MODE (type)) + || (GET_MODE_PRECISION (TYPE_MODE (type)) + == TYPE_PRECISION (type))); + /* fall through */ + + case LSHIFT_EXPR: + case RSHIFT_EXPR: + /* If this is a fixed-point operation, then we cannot use the code + below because "expand_shift" doesn't support sat/no-sat fixed-point + shifts. */ + if (ALL_FIXED_POINT_MODE_P (mode)) + goto binop; + + if (! safe_from_p (subtarget, treeop1, 1)) + subtarget = 0; + if (modifier == EXPAND_STACK_PARM) + target = 0; + op0 = expand_expr (treeop0, subtarget, + VOIDmode, EXPAND_NORMAL); + temp = expand_shift (code, mode, op0, treeop1, target, + unsignedp); + if (code == LSHIFT_EXPR) + temp = REDUCE_BIT_FIELD (temp); + return temp; + + /* Could determine the answer when only additive constants differ. Also, + the addition of one can be handled by changing the condition. */ + case LT_EXPR: + case LE_EXPR: + case GT_EXPR: + case GE_EXPR: + case EQ_EXPR: + case NE_EXPR: + case UNORDERED_EXPR: + case ORDERED_EXPR: + case UNLT_EXPR: + case UNLE_EXPR: + case UNGT_EXPR: + case UNGE_EXPR: + case UNEQ_EXPR: + case LTGT_EXPR: + temp = do_store_flag (ops, + modifier != EXPAND_STACK_PARM ? target : NULL_RTX, + tmode != VOIDmode ? tmode : mode); + if (temp) + return temp; + + /* Use a compare and a jump for BLKmode comparisons, or for function + type comparisons is HAVE_canonicalize_funcptr_for_compare. */ + + if ((target == 0 + || modifier == EXPAND_STACK_PARM + || ! safe_from_p (target, treeop0, 1) + || ! safe_from_p (target, treeop1, 1) + /* Make sure we don't have a hard reg (such as function's return + value) live across basic blocks, if not optimizing. */ + || (!optimize && REG_P (target) + && REGNO (target) < FIRST_PSEUDO_REGISTER))) + target = gen_reg_rtx (tmode != VOIDmode ? tmode : mode); + + emit_move_insn (target, const0_rtx); + + op1 = gen_label_rtx (); + jumpifnot_1 (code, treeop0, treeop1, op1); + + emit_move_insn (target, const1_rtx); + + emit_label (op1); + return target; + + case TRUTH_NOT_EXPR: + if (modifier == EXPAND_STACK_PARM) + target = 0; + op0 = expand_expr (treeop0, target, + VOIDmode, EXPAND_NORMAL); + /* The parser is careful to generate TRUTH_NOT_EXPR + only with operands that are always zero or one. */ + temp = expand_binop (mode, xor_optab, op0, const1_rtx, + target, 1, OPTAB_LIB_WIDEN); + gcc_assert (temp); + return temp; + + case COMPLEX_EXPR: + /* Get the rtx code of the operands. */ + op0 = expand_normal (treeop0); + op1 = expand_normal (treeop1); + + if (!target) + target = gen_reg_rtx (TYPE_MODE (type)); + + /* Move the real (op0) and imaginary (op1) parts to their location. */ + write_complex_part (target, op0, false); + write_complex_part (target, op1, true); + + return target; + + case WIDEN_SUM_EXPR: + { + tree oprnd0 = treeop0; + tree oprnd1 = treeop1; + + expand_operands (oprnd0, oprnd1, NULL_RTX, &op0, &op1, EXPAND_NORMAL); + target = expand_widen_pattern_expr (ops, op0, NULL_RTX, op1, + target, unsignedp); + return target; + } + + case REDUC_MAX_EXPR: + case REDUC_MIN_EXPR: + case REDUC_PLUS_EXPR: + { + op0 = expand_normal (treeop0); + this_optab = optab_for_tree_code (code, type, optab_default); + temp = expand_unop (mode, this_optab, op0, target, unsignedp); + gcc_assert (temp); + return temp; + } + + case VEC_EXTRACT_EVEN_EXPR: + case VEC_EXTRACT_ODD_EXPR: + { + expand_operands (treeop0, treeop1, + NULL_RTX, &op0, &op1, EXPAND_NORMAL); + this_optab = optab_for_tree_code (code, type, optab_default); + temp = expand_binop (mode, this_optab, op0, op1, target, unsignedp, + OPTAB_WIDEN); + gcc_assert (temp); + return temp; + } + + case VEC_INTERLEAVE_HIGH_EXPR: + case VEC_INTERLEAVE_LOW_EXPR: + { + expand_operands (treeop0, treeop1, + NULL_RTX, &op0, &op1, EXPAND_NORMAL); + this_optab = optab_for_tree_code (code, type, optab_default); + temp = expand_binop (mode, this_optab, op0, op1, target, unsignedp, + OPTAB_WIDEN); + gcc_assert (temp); + return temp; + } + + case VEC_LSHIFT_EXPR: + case VEC_RSHIFT_EXPR: + { + target = expand_vec_shift_expr (ops, target); + return target; + } + + case VEC_UNPACK_HI_EXPR: + case VEC_UNPACK_LO_EXPR: + { + op0 = expand_normal (treeop0); + this_optab = optab_for_tree_code (code, type, optab_default); + temp = expand_widen_pattern_expr (ops, op0, NULL_RTX, NULL_RTX, + target, unsignedp); + gcc_assert (temp); + return temp; + } + + case VEC_UNPACK_FLOAT_HI_EXPR: + case VEC_UNPACK_FLOAT_LO_EXPR: + { + op0 = expand_normal (treeop0); + /* The signedness is determined from input operand. */ + this_optab = optab_for_tree_code (code, + TREE_TYPE (treeop0), + optab_default); + temp = expand_widen_pattern_expr + (ops, op0, NULL_RTX, NULL_RTX, + target, TYPE_UNSIGNED (TREE_TYPE (treeop0))); + + gcc_assert (temp); + return temp; + } + + case VEC_WIDEN_MULT_HI_EXPR: + case VEC_WIDEN_MULT_LO_EXPR: + { + tree oprnd0 = treeop0; + tree oprnd1 = treeop1; + + expand_operands (oprnd0, oprnd1, NULL_RTX, &op0, &op1, EXPAND_NORMAL); + target = expand_widen_pattern_expr (ops, op0, op1, NULL_RTX, + target, unsignedp); + gcc_assert (target); + return target; + } + + case VEC_PACK_TRUNC_EXPR: + case VEC_PACK_SAT_EXPR: + case VEC_PACK_FIX_TRUNC_EXPR: + mode = TYPE_MODE (TREE_TYPE (treeop0)); + goto binop; + + default: + gcc_unreachable (); + } + + /* Here to do an ordinary binary operator. */ + binop: + expand_operands (treeop0, treeop1, + subtarget, &op0, &op1, EXPAND_NORMAL); + binop2: + this_optab = optab_for_tree_code (code, type, optab_default); + binop3: + if (modifier == EXPAND_STACK_PARM) + target = 0; + temp = expand_binop (mode, this_optab, op0, op1, target, + unsignedp, OPTAB_LIB_WIDEN); + gcc_assert (temp); + return REDUCE_BIT_FIELD (temp); +} +#undef REDUCE_BIT_FIELD + +static rtx +expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, + enum expand_modifier modifier, rtx *alt_rtl) +{ + rtx op0, op1, temp, decl_rtl; + tree type; + int unsignedp; + enum machine_mode mode; + enum tree_code code = TREE_CODE (exp); + optab this_optab; + rtx subtarget, original_target; + int ignore; + tree context; + bool reduce_bit_field; + location_t loc = EXPR_LOCATION (exp); + struct separate_ops ops; + tree treeop0, treeop1, treeop2; + type = TREE_TYPE (exp); mode = TYPE_MODE (type); unsignedp = TYPE_UNSIGNED (type); @@ -8177,111 +9218,6 @@ expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, } return expand_call (exp, target, ignore); - case PAREN_EXPR: - CASE_CONVERT: - if (treeop0 == error_mark_node) - return const0_rtx; - - if (TREE_CODE (type) == UNION_TYPE) - { - tree valtype = TREE_TYPE (treeop0); - - /* If both input and output are BLKmode, this conversion isn't doing - anything except possibly changing memory attribute. */ - if (mode == BLKmode && TYPE_MODE (valtype) == BLKmode) - { - rtx result = expand_expr (treeop0, target, tmode, - modifier); - - result = copy_rtx (result); - set_mem_attributes (result, exp, 0); - return result; - } - - if (target == 0) - { - if (TYPE_MODE (type) != BLKmode) - target = gen_reg_rtx (TYPE_MODE (type)); - else - target = assign_temp (type, 0, 1, 1); - } - - if (MEM_P (target)) - /* Store data into beginning of memory target. */ - store_expr (treeop0, - adjust_address (target, TYPE_MODE (valtype), 0), - modifier == EXPAND_STACK_PARM, - false); - - else - { - gcc_assert (REG_P (target)); - - /* Store this field into a union of the proper type. */ - store_field (target, - MIN ((int_size_in_bytes (TREE_TYPE - (treeop0)) - * BITS_PER_UNIT), - (HOST_WIDE_INT) GET_MODE_BITSIZE (mode)), - 0, TYPE_MODE (valtype), treeop0, - type, 0, false); - } - - /* Return the entire union. */ - return target; - } - - if (mode == TYPE_MODE (TREE_TYPE (treeop0))) - { - op0 = expand_expr (treeop0, target, VOIDmode, - modifier); - - /* If the signedness of the conversion differs and OP0 is - a promoted SUBREG, clear that indication since we now - have to do the proper extension. */ - if (TYPE_UNSIGNED (TREE_TYPE (treeop0)) != unsignedp - && GET_CODE (op0) == SUBREG) - SUBREG_PROMOTED_VAR_P (op0) = 0; - - return REDUCE_BIT_FIELD (op0); - } - - op0 = expand_expr (treeop0, NULL_RTX, mode, - modifier == EXPAND_SUM ? EXPAND_NORMAL : modifier); - if (GET_MODE (op0) == mode) - ; - - /* If OP0 is a constant, just convert it into the proper mode. */ - else if (CONSTANT_P (op0)) - { - tree inner_type = TREE_TYPE (treeop0); - enum machine_mode inner_mode = TYPE_MODE (inner_type); - - if (modifier == EXPAND_INITIALIZER) - op0 = simplify_gen_subreg (mode, op0, inner_mode, - subreg_lowpart_offset (mode, - inner_mode)); - else - op0= convert_modes (mode, inner_mode, op0, - TYPE_UNSIGNED (inner_type)); - } - - else if (modifier == EXPAND_INITIALIZER) - op0 = gen_rtx_fmt_e (unsignedp ? ZERO_EXTEND : SIGN_EXTEND, mode, op0); - - else if (target == 0) - op0 = convert_to_mode (mode, op0, - TYPE_UNSIGNED (TREE_TYPE - (treeop0))); - else - { - convert_move (target, op0, - TYPE_UNSIGNED (TREE_TYPE (treeop0))); - op0 = target; - } - - return REDUCE_BIT_FIELD (op0); - case VIEW_CONVERT_EXPR: op0 = NULL_RTX; @@ -8432,720 +9368,6 @@ expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, return op0; - case POINTER_PLUS_EXPR: - /* Even though the sizetype mode and the pointer's mode can be different - expand is able to handle this correctly and get the correct result out - of the PLUS_EXPR code. */ - /* Make sure to sign-extend the sizetype offset in a POINTER_PLUS_EXPR - if sizetype precision is smaller than pointer precision. */ - if (TYPE_PRECISION (sizetype) < TYPE_PRECISION (type)) - treeop1 = fold_convert_loc (loc, type, - fold_convert_loc (loc, ssizetype, - treeop1)); - case PLUS_EXPR: - - /* Check if this is a case for multiplication and addition. */ - if ((TREE_CODE (type) == INTEGER_TYPE - || TREE_CODE (type) == FIXED_POINT_TYPE) - && (subexp0_def = get_def_for_expr (treeop0, - MULT_EXPR))) - { - tree subsubexp0, subsubexp1; - gimple subsubexp0_def, subsubexp1_def; - enum tree_code this_code; - - this_code = TREE_CODE (type) == INTEGER_TYPE ? NOP_EXPR - : FIXED_CONVERT_EXPR; - subsubexp0 = gimple_assign_rhs1 (subexp0_def); - subsubexp0_def = get_def_for_expr (subsubexp0, this_code); - subsubexp1 = gimple_assign_rhs2 (subexp0_def); - subsubexp1_def = get_def_for_expr (subsubexp1, this_code); - if (subsubexp0_def && subsubexp1_def - && (top0 = gimple_assign_rhs1 (subsubexp0_def)) - && (top1 = gimple_assign_rhs1 (subsubexp1_def)) - && (TYPE_PRECISION (TREE_TYPE (top0)) - < TYPE_PRECISION (TREE_TYPE (subsubexp0))) - && (TYPE_PRECISION (TREE_TYPE (top0)) - == TYPE_PRECISION (TREE_TYPE (top1))) - && (TYPE_UNSIGNED (TREE_TYPE (top0)) - == TYPE_UNSIGNED (TREE_TYPE (top1)))) - { - tree op0type = TREE_TYPE (top0); - enum machine_mode innermode = TYPE_MODE (op0type); - bool zextend_p = TYPE_UNSIGNED (op0type); - bool sat_p = TYPE_SATURATING (TREE_TYPE (subsubexp0)); - if (sat_p == 0) - this_optab = zextend_p ? umadd_widen_optab : smadd_widen_optab; - else - this_optab = zextend_p ? usmadd_widen_optab - : ssmadd_widen_optab; - if (mode == GET_MODE_2XWIDER_MODE (innermode) - && (optab_handler (this_optab, mode)->insn_code - != CODE_FOR_nothing)) - { - expand_operands (top0, top1, NULL_RTX, &op0, &op1, - EXPAND_NORMAL); - op2 = expand_expr (treeop1, subtarget, - VOIDmode, EXPAND_NORMAL); - temp = expand_ternary_op (mode, this_optab, op0, op1, op2, - target, unsignedp); - gcc_assert (temp); - return REDUCE_BIT_FIELD (temp); - } - } - } - - /* If we are adding a constant, a VAR_DECL that is sp, fp, or ap, and - something else, make sure we add the register to the constant and - then to the other thing. This case can occur during strength - reduction and doing it this way will produce better code if the - frame pointer or argument pointer is eliminated. - - fold-const.c will ensure that the constant is always in the inner - PLUS_EXPR, so the only case we need to do anything about is if - sp, ap, or fp is our second argument, in which case we must swap - the innermost first argument and our second argument. */ - - if (TREE_CODE (treeop0) == PLUS_EXPR - && TREE_CODE (TREE_OPERAND (treeop0, 1)) == INTEGER_CST - && TREE_CODE (treeop1) == VAR_DECL - && (DECL_RTL (treeop1) == frame_pointer_rtx - || DECL_RTL (treeop1) == stack_pointer_rtx - || DECL_RTL (treeop1) == arg_pointer_rtx)) - { - tree t = treeop1; - - treeop1 = TREE_OPERAND (treeop0, 0); - TREE_OPERAND (treeop0, 0) = t; - } - - /* If the result is to be ptr_mode and we are adding an integer to - something, we might be forming a constant. So try to use - plus_constant. If it produces a sum and we can't accept it, - use force_operand. This allows P = &ARR[const] to generate - efficient code on machines where a SYMBOL_REF is not a valid - address. - - If this is an EXPAND_SUM call, always return the sum. */ - if (modifier == EXPAND_SUM || modifier == EXPAND_INITIALIZER - || (mode == ptr_mode && (unsignedp || ! flag_trapv))) - { - if (modifier == EXPAND_STACK_PARM) - target = 0; - if (TREE_CODE (treeop0) == INTEGER_CST - && GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_WIDE_INT - && TREE_CONSTANT (treeop1)) - { - rtx constant_part; - - op1 = expand_expr (treeop1, subtarget, VOIDmode, - EXPAND_SUM); - /* Use immed_double_const to ensure that the constant is - truncated according to the mode of OP1, then sign extended - to a HOST_WIDE_INT. Using the constant directly can result - in non-canonical RTL in a 64x32 cross compile. */ - constant_part - = immed_double_const (TREE_INT_CST_LOW (treeop0), - (HOST_WIDE_INT) 0, - TYPE_MODE (TREE_TYPE (treeop1))); - op1 = plus_constant (op1, INTVAL (constant_part)); - if (modifier != EXPAND_SUM && modifier != EXPAND_INITIALIZER) - op1 = force_operand (op1, target); - return REDUCE_BIT_FIELD (op1); - } - - else if (TREE_CODE (treeop1) == INTEGER_CST - && GET_MODE_BITSIZE (mode) <= HOST_BITS_PER_WIDE_INT - && TREE_CONSTANT (treeop0)) - { - rtx constant_part; - - op0 = expand_expr (treeop0, subtarget, VOIDmode, - (modifier == EXPAND_INITIALIZER - ? EXPAND_INITIALIZER : EXPAND_SUM)); - if (! CONSTANT_P (op0)) - { - op1 = expand_expr (treeop1, NULL_RTX, - VOIDmode, modifier); - /* Return a PLUS if modifier says it's OK. */ - if (modifier == EXPAND_SUM - || modifier == EXPAND_INITIALIZER) - return simplify_gen_binary (PLUS, mode, op0, op1); - goto binop2; - } - /* Use immed_double_const to ensure that the constant is - truncated according to the mode of OP1, then sign extended - to a HOST_WIDE_INT. Using the constant directly can result - in non-canonical RTL in a 64x32 cross compile. */ - constant_part - = immed_double_const (TREE_INT_CST_LOW (treeop1), - (HOST_WIDE_INT) 0, - TYPE_MODE (TREE_TYPE (treeop0))); - op0 = plus_constant (op0, INTVAL (constant_part)); - if (modifier != EXPAND_SUM && modifier != EXPAND_INITIALIZER) - op0 = force_operand (op0, target); - return REDUCE_BIT_FIELD (op0); - } - } - - /* No sense saving up arithmetic to be done - if it's all in the wrong mode to form part of an address. - And force_operand won't know whether to sign-extend or - zero-extend. */ - if ((modifier != EXPAND_SUM && modifier != EXPAND_INITIALIZER) - || mode != ptr_mode) - { - expand_operands (treeop0, treeop1, - subtarget, &op0, &op1, EXPAND_NORMAL); - if (op0 == const0_rtx) - return op1; - if (op1 == const0_rtx) - return op0; - goto binop2; - } - - expand_operands (treeop0, treeop1, - subtarget, &op0, &op1, modifier); - return REDUCE_BIT_FIELD (simplify_gen_binary (PLUS, mode, op0, op1)); - - case MINUS_EXPR: - /* Check if this is a case for multiplication and subtraction. */ - if ((TREE_CODE (type) == INTEGER_TYPE - || TREE_CODE (type) == FIXED_POINT_TYPE) - && (subexp1_def = get_def_for_expr (treeop1, - MULT_EXPR))) - { - tree subsubexp0, subsubexp1; - gimple subsubexp0_def, subsubexp1_def; - enum tree_code this_code; - - this_code = TREE_CODE (type) == INTEGER_TYPE ? NOP_EXPR - : FIXED_CONVERT_EXPR; - subsubexp0 = gimple_assign_rhs1 (subexp1_def); - subsubexp0_def = get_def_for_expr (subsubexp0, this_code); - subsubexp1 = gimple_assign_rhs2 (subexp1_def); - subsubexp1_def = get_def_for_expr (subsubexp1, this_code); - if (subsubexp0_def && subsubexp1_def - && (top0 = gimple_assign_rhs1 (subsubexp0_def)) - && (top1 = gimple_assign_rhs1 (subsubexp1_def)) - && (TYPE_PRECISION (TREE_TYPE (top0)) - < TYPE_PRECISION (TREE_TYPE (subsubexp0))) - && (TYPE_PRECISION (TREE_TYPE (top0)) - == TYPE_PRECISION (TREE_TYPE (top1))) - && (TYPE_UNSIGNED (TREE_TYPE (top0)) - == TYPE_UNSIGNED (TREE_TYPE (top1)))) - { - tree op0type = TREE_TYPE (top0); - enum machine_mode innermode = TYPE_MODE (op0type); - bool zextend_p = TYPE_UNSIGNED (op0type); - bool sat_p = TYPE_SATURATING (TREE_TYPE (subsubexp0)); - if (sat_p == 0) - this_optab = zextend_p ? umsub_widen_optab : smsub_widen_optab; - else - this_optab = zextend_p ? usmsub_widen_optab - : ssmsub_widen_optab; - if (mode == GET_MODE_2XWIDER_MODE (innermode) - && (optab_handler (this_optab, mode)->insn_code - != CODE_FOR_nothing)) - { - expand_operands (top0, top1, NULL_RTX, &op0, &op1, - EXPAND_NORMAL); - op2 = expand_expr (treeop0, subtarget, - VOIDmode, EXPAND_NORMAL); - temp = expand_ternary_op (mode, this_optab, op0, op1, op2, - target, unsignedp); - gcc_assert (temp); - return REDUCE_BIT_FIELD (temp); - } - } - } - - /* For initializers, we are allowed to return a MINUS of two - symbolic constants. Here we handle all cases when both operands - are constant. */ - /* Handle difference of two symbolic constants, - for the sake of an initializer. */ - if ((modifier == EXPAND_SUM || modifier == EXPAND_INITIALIZER) - && really_constant_p (treeop0) - && really_constant_p (treeop1)) - { - expand_operands (treeop0, treeop1, - NULL_RTX, &op0, &op1, modifier); - - /* If the last operand is a CONST_INT, use plus_constant of - the negated constant. Else make the MINUS. */ - if (CONST_INT_P (op1)) - return REDUCE_BIT_FIELD (plus_constant (op0, - INTVAL (op1))); - else - return REDUCE_BIT_FIELD (gen_rtx_MINUS (mode, op0, op1)); - } - - /* No sense saving up arithmetic to be done - if it's all in the wrong mode to form part of an address. - And force_operand won't know whether to sign-extend or - zero-extend. */ - if ((modifier != EXPAND_SUM && modifier != EXPAND_INITIALIZER) - || mode != ptr_mode) - goto binop; - - expand_operands (treeop0, treeop1, - subtarget, &op0, &op1, modifier); - - /* Convert A - const to A + (-const). */ - if (CONST_INT_P (op1)) - { - op1 = negate_rtx (mode, op1); - return REDUCE_BIT_FIELD (simplify_gen_binary (PLUS, mode, op0, op1)); - } - - goto binop2; - - case MULT_EXPR: - /* If this is a fixed-point operation, then we cannot use the code - below because "expand_mult" doesn't support sat/no-sat fixed-point - multiplications. */ - if (ALL_FIXED_POINT_MODE_P (mode)) - goto binop; - - /* If first operand is constant, swap them. - Thus the following special case checks need only - check the second operand. */ - if (TREE_CODE (treeop0) == INTEGER_CST) - { - tree t1 = treeop0; - treeop0 = treeop1; - treeop1 = t1; - } - - /* Attempt to return something suitable for generating an - indexed address, for machines that support that. */ - - if (modifier == EXPAND_SUM && mode == ptr_mode - && host_integerp (treeop1, 0)) - { - tree exp1 = treeop1; - - op0 = expand_expr (treeop0, subtarget, VOIDmode, - EXPAND_SUM); - - if (!REG_P (op0)) - op0 = force_operand (op0, NULL_RTX); - if (!REG_P (op0)) - op0 = copy_to_mode_reg (mode, op0); - - return REDUCE_BIT_FIELD (gen_rtx_MULT (mode, op0, - gen_int_mode (tree_low_cst (exp1, 0), - TYPE_MODE (TREE_TYPE (exp1))))); - } - - if (modifier == EXPAND_STACK_PARM) - target = 0; - - /* Check for multiplying things that have been extended - from a narrower type. If this machine supports multiplying - in that narrower type with a result in the desired type, - do it that way, and avoid the explicit type-conversion. */ - - subexp0 = treeop0; - subexp1 = treeop1; - subexp0_def = get_def_for_expr (subexp0, NOP_EXPR); - subexp1_def = get_def_for_expr (subexp1, NOP_EXPR); - top0 = top1 = NULL_TREE; - - /* First, check if we have a multiplication of one signed and one - unsigned operand. */ - if (subexp0_def - && (top0 = gimple_assign_rhs1 (subexp0_def)) - && subexp1_def - && (top1 = gimple_assign_rhs1 (subexp1_def)) - && TREE_CODE (type) == INTEGER_TYPE - && (TYPE_PRECISION (TREE_TYPE (top0)) - < TYPE_PRECISION (TREE_TYPE (subexp0))) - && (TYPE_PRECISION (TREE_TYPE (top0)) - == TYPE_PRECISION (TREE_TYPE (top1))) - && (TYPE_UNSIGNED (TREE_TYPE (top0)) - != TYPE_UNSIGNED (TREE_TYPE (top1)))) - { - enum machine_mode innermode - = TYPE_MODE (TREE_TYPE (top0)); - this_optab = usmul_widen_optab; - if (mode == GET_MODE_WIDER_MODE (innermode)) - { - if (optab_handler (this_optab, mode)->insn_code != CODE_FOR_nothing) - { - if (TYPE_UNSIGNED (TREE_TYPE (top0))) - expand_operands (top0, top1, NULL_RTX, &op0, &op1, - EXPAND_NORMAL); - else - expand_operands (top0, top1, NULL_RTX, &op1, &op0, - EXPAND_NORMAL); - - goto binop3; - } - } - } - /* Check for a multiplication with matching signedness. If - valid, TOP0 and TOP1 were set in the previous if - condition. */ - else if (top0 - && TREE_CODE (type) == INTEGER_TYPE - && (TYPE_PRECISION (TREE_TYPE (top0)) - < TYPE_PRECISION (TREE_TYPE (subexp0))) - && ((TREE_CODE (subexp1) == INTEGER_CST - && int_fits_type_p (subexp1, TREE_TYPE (top0)) - /* Don't use a widening multiply if a shift will do. */ - && ((GET_MODE_BITSIZE (TYPE_MODE (TREE_TYPE (subexp1))) - > HOST_BITS_PER_WIDE_INT) - || exact_log2 (TREE_INT_CST_LOW (subexp1)) < 0)) - || - (top1 - && (TYPE_PRECISION (TREE_TYPE (top1)) - == TYPE_PRECISION (TREE_TYPE (top0)) - /* If both operands are extended, they must either both - be zero-extended or both be sign-extended. */ - && (TYPE_UNSIGNED (TREE_TYPE (top1)) - == TYPE_UNSIGNED (TREE_TYPE (top0))))))) - { - tree op0type = TREE_TYPE (top0); - enum machine_mode innermode = TYPE_MODE (op0type); - bool zextend_p = TYPE_UNSIGNED (op0type); - optab other_optab = zextend_p ? smul_widen_optab : umul_widen_optab; - this_optab = zextend_p ? umul_widen_optab : smul_widen_optab; - - if (mode == GET_MODE_2XWIDER_MODE (innermode)) - { - if (optab_handler (this_optab, mode)->insn_code != CODE_FOR_nothing) - { - if (TREE_CODE (subexp1) == INTEGER_CST) - expand_operands (top0, subexp1, NULL_RTX, &op0, &op1, - EXPAND_NORMAL); - else - expand_operands (top0, top1, NULL_RTX, &op0, &op1, - EXPAND_NORMAL); - goto binop3; - } - else if (optab_handler (other_optab, mode)->insn_code != CODE_FOR_nothing - && innermode == word_mode) - { - rtx htem, hipart; - op0 = expand_normal (top0); - if (TREE_CODE (subexp1) == INTEGER_CST) - op1 = convert_modes (innermode, mode, - expand_normal (subexp1), unsignedp); - else - op1 = expand_normal (top1); - temp = expand_binop (mode, other_optab, op0, op1, target, - unsignedp, OPTAB_LIB_WIDEN); - hipart = gen_highpart (innermode, temp); - htem = expand_mult_highpart_adjust (innermode, hipart, - op0, op1, hipart, - zextend_p); - if (htem != hipart) - emit_move_insn (hipart, htem); - return REDUCE_BIT_FIELD (temp); - } - } - } - expand_operands (subexp0, subexp1, subtarget, &op0, &op1, EXPAND_NORMAL); - return REDUCE_BIT_FIELD (expand_mult (mode, op0, op1, target, unsignedp)); - - case TRUNC_DIV_EXPR: - case FLOOR_DIV_EXPR: - case CEIL_DIV_EXPR: - case ROUND_DIV_EXPR: - case EXACT_DIV_EXPR: - /* If this is a fixed-point operation, then we cannot use the code - below because "expand_divmod" doesn't support sat/no-sat fixed-point - divisions. */ - if (ALL_FIXED_POINT_MODE_P (mode)) - goto binop; - - if (modifier == EXPAND_STACK_PARM) - target = 0; - /* Possible optimization: compute the dividend with EXPAND_SUM - then if the divisor is constant can optimize the case - where some terms of the dividend have coeffs divisible by it. */ - expand_operands (treeop0, treeop1, - subtarget, &op0, &op1, EXPAND_NORMAL); - return expand_divmod (0, code, mode, op0, op1, target, unsignedp); - - case RDIV_EXPR: - goto binop; - - case TRUNC_MOD_EXPR: - case FLOOR_MOD_EXPR: - case CEIL_MOD_EXPR: - case ROUND_MOD_EXPR: - if (modifier == EXPAND_STACK_PARM) - target = 0; - expand_operands (treeop0, treeop1, - subtarget, &op0, &op1, EXPAND_NORMAL); - return expand_divmod (1, code, mode, op0, op1, target, unsignedp); - - case FIXED_CONVERT_EXPR: - op0 = expand_normal (treeop0); - if (target == 0 || modifier == EXPAND_STACK_PARM) - target = gen_reg_rtx (mode); - - if ((TREE_CODE (TREE_TYPE (treeop0)) == INTEGER_TYPE - && TYPE_UNSIGNED (TREE_TYPE (treeop0))) - || (TREE_CODE (type) == INTEGER_TYPE && TYPE_UNSIGNED (type))) - expand_fixed_convert (target, op0, 1, TYPE_SATURATING (type)); - else - expand_fixed_convert (target, op0, 0, TYPE_SATURATING (type)); - return target; - - case FIX_TRUNC_EXPR: - op0 = expand_normal (treeop0); - if (target == 0 || modifier == EXPAND_STACK_PARM) - target = gen_reg_rtx (mode); - expand_fix (target, op0, unsignedp); - return target; - - case FLOAT_EXPR: - op0 = expand_normal (treeop0); - if (target == 0 || modifier == EXPAND_STACK_PARM) - target = gen_reg_rtx (mode); - /* expand_float can't figure out what to do if FROM has VOIDmode. - So give it the correct mode. With -O, cse will optimize this. */ - if (GET_MODE (op0) == VOIDmode) - op0 = copy_to_mode_reg (TYPE_MODE (TREE_TYPE (treeop0)), - op0); - expand_float (target, op0, - TYPE_UNSIGNED (TREE_TYPE (treeop0))); - return target; - - case NEGATE_EXPR: - op0 = expand_expr (treeop0, subtarget, - VOIDmode, EXPAND_NORMAL); - if (modifier == EXPAND_STACK_PARM) - target = 0; - temp = expand_unop (mode, - optab_for_tree_code (NEGATE_EXPR, type, - optab_default), - op0, target, 0); - gcc_assert (temp); - return REDUCE_BIT_FIELD (temp); - - case ABS_EXPR: - op0 = expand_expr (treeop0, subtarget, - VOIDmode, EXPAND_NORMAL); - if (modifier == EXPAND_STACK_PARM) - target = 0; - - /* ABS_EXPR is not valid for complex arguments. */ - gcc_assert (GET_MODE_CLASS (mode) != MODE_COMPLEX_INT - && GET_MODE_CLASS (mode) != MODE_COMPLEX_FLOAT); - - /* Unsigned abs is simply the operand. Testing here means we don't - risk generating incorrect code below. */ - if (TYPE_UNSIGNED (type)) - return op0; - - return expand_abs (mode, op0, target, unsignedp, - safe_from_p (target, treeop0, 1)); - - case MAX_EXPR: - case MIN_EXPR: - target = original_target; - if (target == 0 - || modifier == EXPAND_STACK_PARM - || (MEM_P (target) && MEM_VOLATILE_P (target)) - || GET_MODE (target) != mode - || (REG_P (target) - && REGNO (target) < FIRST_PSEUDO_REGISTER)) - target = gen_reg_rtx (mode); - expand_operands (treeop0, treeop1, - target, &op0, &op1, EXPAND_NORMAL); - - /* First try to do it with a special MIN or MAX instruction. - If that does not win, use a conditional jump to select the proper - value. */ - this_optab = optab_for_tree_code (code, type, optab_default); - temp = expand_binop (mode, this_optab, op0, op1, target, unsignedp, - OPTAB_WIDEN); - if (temp != 0) - return temp; - - /* At this point, a MEM target is no longer useful; we will get better - code without it. */ - - if (! REG_P (target)) - target = gen_reg_rtx (mode); - - /* If op1 was placed in target, swap op0 and op1. */ - if (target != op0 && target == op1) - { - temp = op0; - op0 = op1; - op1 = temp; - } - - /* We generate better code and avoid problems with op1 mentioning - target by forcing op1 into a pseudo if it isn't a constant. */ - if (! CONSTANT_P (op1)) - op1 = force_reg (mode, op1); - - { - enum rtx_code comparison_code; - rtx cmpop1 = op1; - - if (code == MAX_EXPR) - comparison_code = unsignedp ? GEU : GE; - else - comparison_code = unsignedp ? LEU : LE; - - /* Canonicalize to comparisons against 0. */ - if (op1 == const1_rtx) - { - /* Converting (a >= 1 ? a : 1) into (a > 0 ? a : 1) - or (a != 0 ? a : 1) for unsigned. - For MIN we are safe converting (a <= 1 ? a : 1) - into (a <= 0 ? a : 1) */ - cmpop1 = const0_rtx; - if (code == MAX_EXPR) - comparison_code = unsignedp ? NE : GT; - } - if (op1 == constm1_rtx && !unsignedp) - { - /* Converting (a >= -1 ? a : -1) into (a >= 0 ? a : -1) - and (a <= -1 ? a : -1) into (a < 0 ? a : -1) */ - cmpop1 = const0_rtx; - if (code == MIN_EXPR) - comparison_code = LT; - } -#ifdef HAVE_conditional_move - /* Use a conditional move if possible. */ - if (can_conditionally_move_p (mode)) - { - rtx insn; - - /* ??? Same problem as in expmed.c: emit_conditional_move - forces a stack adjustment via compare_from_rtx, and we - lose the stack adjustment if the sequence we are about - to create is discarded. */ - do_pending_stack_adjust (); - - start_sequence (); - - /* Try to emit the conditional move. */ - insn = emit_conditional_move (target, comparison_code, - op0, cmpop1, mode, - op0, op1, mode, - unsignedp); - - /* If we could do the conditional move, emit the sequence, - and return. */ - if (insn) - { - rtx seq = get_insns (); - end_sequence (); - emit_insn (seq); - return target; - } - - /* Otherwise discard the sequence and fall back to code with - branches. */ - end_sequence (); - } -#endif - if (target != op0) - emit_move_insn (target, op0); - - temp = gen_label_rtx (); - do_compare_rtx_and_jump (target, cmpop1, comparison_code, - unsignedp, mode, NULL_RTX, NULL_RTX, temp); - } - emit_move_insn (target, op1); - emit_label (temp); - return target; - - case BIT_NOT_EXPR: - op0 = expand_expr (treeop0, subtarget, - VOIDmode, EXPAND_NORMAL); - if (modifier == EXPAND_STACK_PARM) - target = 0; - temp = expand_unop (mode, one_cmpl_optab, op0, target, 1); - gcc_assert (temp); - return temp; - - /* ??? Can optimize bitwise operations with one arg constant. - Can optimize (a bitwise1 n) bitwise2 (a bitwise3 b) - and (a bitwise1 b) bitwise2 b (etc) - but that is probably not worth while. */ - - /* BIT_AND_EXPR is for bitwise anding. TRUTH_AND_EXPR is for anding two - boolean values when we want in all cases to compute both of them. In - general it is fastest to do TRUTH_AND_EXPR by computing both operands - as actual zero-or-1 values and then bitwise anding. In cases where - there cannot be any side effects, better code would be made by - treating TRUTH_AND_EXPR like TRUTH_ANDIF_EXPR; but the question is - how to recognize those cases. */ - - case TRUTH_AND_EXPR: - code = BIT_AND_EXPR; - case BIT_AND_EXPR: - goto binop; - - case TRUTH_OR_EXPR: - code = BIT_IOR_EXPR; - case BIT_IOR_EXPR: - goto binop; - - case TRUTH_XOR_EXPR: - code = BIT_XOR_EXPR; - case BIT_XOR_EXPR: - goto binop; - - case LROTATE_EXPR: - case RROTATE_EXPR: - gcc_assert (VECTOR_MODE_P (TYPE_MODE (type)) - || (GET_MODE_PRECISION (TYPE_MODE (type)) - == TYPE_PRECISION (type))); - /* fall through */ - - case LSHIFT_EXPR: - case RSHIFT_EXPR: - /* If this is a fixed-point operation, then we cannot use the code - below because "expand_shift" doesn't support sat/no-sat fixed-point - shifts. */ - if (ALL_FIXED_POINT_MODE_P (mode)) - goto binop; - - if (! safe_from_p (subtarget, treeop1, 1)) - subtarget = 0; - if (modifier == EXPAND_STACK_PARM) - target = 0; - op0 = expand_expr (treeop0, subtarget, - VOIDmode, EXPAND_NORMAL); - temp = expand_shift (code, mode, op0, treeop1, target, - unsignedp); - if (code == LSHIFT_EXPR) - temp = REDUCE_BIT_FIELD (temp); - return temp; - - /* Could determine the answer when only additive constants differ. Also, - the addition of one can be handled by changing the condition. */ - case LT_EXPR: - case LE_EXPR: - case GT_EXPR: - case GE_EXPR: - case EQ_EXPR: - case NE_EXPR: - case UNORDERED_EXPR: - case ORDERED_EXPR: - case UNLT_EXPR: - case UNLE_EXPR: - case UNGT_EXPR: - case UNGE_EXPR: - case UNEQ_EXPR: - case LTGT_EXPR: - temp = do_store_flag (&ops, - modifier != EXPAND_STACK_PARM ? target : NULL_RTX, - tmode != VOIDmode ? tmode : mode); - if (temp) - return temp; - /* Use a compare and a jump for BLKmode comparisons, or for function type comparisons is HAVE_canonicalize_funcptr_for_compare. */ @@ -9176,18 +9398,6 @@ expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, emit_label (op1); return ignore ? const0_rtx : target; - case TRUTH_NOT_EXPR: - if (modifier == EXPAND_STACK_PARM) - target = 0; - op0 = expand_expr (treeop0, target, - VOIDmode, EXPAND_NORMAL); - /* The parser is careful to generate TRUTH_NOT_EXPR - only with operands that are always zero or one. */ - temp = expand_binop (mode, xor_optab, op0, const1_rtx, - target, 1, OPTAB_LIB_WIDEN); - gcc_assert (temp); - return temp; - case STATEMENT_LIST: { tree_stmt_iterator iter; @@ -9305,20 +9515,6 @@ expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, case ADDR_EXPR: return expand_expr_addr_expr (exp, target, tmode, modifier); - case COMPLEX_EXPR: - /* Get the rtx code of the operands. */ - op0 = expand_normal (treeop0); - op1 = expand_normal (treeop1); - - if (!target) - target = gen_reg_rtx (TYPE_MODE (type)); - - /* Move the real (op0) and imaginary (op1) parts to their location. */ - write_complex_part (target, op0, false); - write_complex_part (target, op1, true); - - return target; - case REALPART_EXPR: op0 = expand_normal (treeop0); return read_complex_part (op0, false); @@ -9415,105 +9611,6 @@ expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, return target; } - case WIDEN_SUM_EXPR: - { - tree oprnd0 = treeop0; - tree oprnd1 = treeop1; - - expand_operands (oprnd0, oprnd1, NULL_RTX, &op0, &op1, EXPAND_NORMAL); - target = expand_widen_pattern_expr (&ops, op0, NULL_RTX, op1, - target, unsignedp); - return target; - } - - case REDUC_MAX_EXPR: - case REDUC_MIN_EXPR: - case REDUC_PLUS_EXPR: - { - op0 = expand_normal (treeop0); - this_optab = optab_for_tree_code (code, type, optab_default); - temp = expand_unop (mode, this_optab, op0, target, unsignedp); - gcc_assert (temp); - return temp; - } - - case VEC_EXTRACT_EVEN_EXPR: - case VEC_EXTRACT_ODD_EXPR: - { - expand_operands (treeop0, treeop1, - NULL_RTX, &op0, &op1, EXPAND_NORMAL); - this_optab = optab_for_tree_code (code, type, optab_default); - temp = expand_binop (mode, this_optab, op0, op1, target, unsignedp, - OPTAB_WIDEN); - gcc_assert (temp); - return temp; - } - - case VEC_INTERLEAVE_HIGH_EXPR: - case VEC_INTERLEAVE_LOW_EXPR: - { - expand_operands (treeop0, treeop1, - NULL_RTX, &op0, &op1, EXPAND_NORMAL); - this_optab = optab_for_tree_code (code, type, optab_default); - temp = expand_binop (mode, this_optab, op0, op1, target, unsignedp, - OPTAB_WIDEN); - gcc_assert (temp); - return temp; - } - - case VEC_LSHIFT_EXPR: - case VEC_RSHIFT_EXPR: - { - target = expand_vec_shift_expr (&ops, target); - return target; - } - - case VEC_UNPACK_HI_EXPR: - case VEC_UNPACK_LO_EXPR: - { - op0 = expand_normal (treeop0); - this_optab = optab_for_tree_code (code, type, optab_default); - temp = expand_widen_pattern_expr (&ops, op0, NULL_RTX, NULL_RTX, - target, unsignedp); - gcc_assert (temp); - return temp; - } - - case VEC_UNPACK_FLOAT_HI_EXPR: - case VEC_UNPACK_FLOAT_LO_EXPR: - { - op0 = expand_normal (treeop0); - /* The signedness is determined from input operand. */ - this_optab = optab_for_tree_code (code, - TREE_TYPE (treeop0), - optab_default); - temp = expand_widen_pattern_expr - (&ops, op0, NULL_RTX, NULL_RTX, - target, TYPE_UNSIGNED (TREE_TYPE (treeop0))); - - gcc_assert (temp); - return temp; - } - - case VEC_WIDEN_MULT_HI_EXPR: - case VEC_WIDEN_MULT_LO_EXPR: - { - tree oprnd0 = treeop0; - tree oprnd1 = treeop1; - - expand_operands (oprnd0, oprnd1, NULL_RTX, &op0, &op1, EXPAND_NORMAL); - target = expand_widen_pattern_expr (&ops, op0, op1, NULL_RTX, - target, unsignedp); - gcc_assert (target); - return target; - } - - case VEC_PACK_TRUNC_EXPR: - case VEC_PACK_SAT_EXPR: - case VEC_PACK_FIX_TRUNC_EXPR: - mode = TYPE_MODE (TREE_TYPE (treeop0)); - goto binop; - case COMPOUND_LITERAL_EXPR: { /* Initialize the anonymous variable declared in the compound @@ -9536,24 +9633,9 @@ expand_expr_real_1 (tree exp, rtx target, enum machine_mode tmode, } default: - gcc_unreachable (); + return expand_expr_real_2 (&ops, target, tmode, modifier); } - - /* Here to do an ordinary binary operator. */ - binop: - expand_operands (treeop0, treeop1, - subtarget, &op0, &op1, EXPAND_NORMAL); - binop2: - this_optab = optab_for_tree_code (code, type, optab_default); - binop3: - if (modifier == EXPAND_STACK_PARM) - target = 0; - temp = expand_binop (mode, this_optab, op0, op1, target, - unsignedp, OPTAB_LIB_WIDEN); - gcc_assert (temp); - return REDUCE_BIT_FIELD (temp); } -#undef REDUCE_BIT_FIELD /* Subroutine of above: reduce EXP to the precision of TYPE (in the signedness of TYPE), possibly returning the result in TARGET. */ |