Lines Matching refs:gimple_build_assign
710 gassign *new_stmt2 = gimple_build_assign (new_lhs, NOP_EXPR, new_rhs); in vect_split_statement()
775 gassign *new_stmt = gimple_build_assign (input, NOP_EXPR, in vect_convert_input()
790 gassign *new_stmt = gimple_build_assign (new_op, NOP_EXPR, input); in vect_convert_input()
844 pattern_stmt = gimple_build_assign (cast_var, NOP_EXPR, lhs); in vect_convert_output()
1014 pattern_stmt = gimple_build_assign (var, DOT_PROD_EXPR, in vect_recog_dot_prod_pattern()
1172 gimple *pattern_stmt = gimple_build_assign (var, SAD_EXPR, sad_oprnd[0], in vect_recog_sad_pattern()
1249 gimple *pattern_stmt = gimple_build_assign (var, wide_code, in vect_recog_widen_op_pattern()
1366 gimple *g = gimple_build_assign (def, MULT_EXPR, exp, logc); in vect_recog_pow_pattern()
1392 stmt = gimple_build_assign (var, MULT_EXPR, base, base); in vect_recog_pow_pattern()
1498 pattern_stmt = gimple_build_assign (var, WIDEN_SUM_EXPR, unprom0.op, oprnd1); in vect_recog_widen_sum_pattern()
1725 gimple *pattern_stmt = gimple_build_assign (new_var, code, in vect_recog_over_widening_pattern()
2082 g = gimple_build_assign (shifted_op0, RSHIFT_EXPR, new_ops[0], one_cst); in vect_recog_average_pattern()
2086 g = gimple_build_assign (shifted_op1, RSHIFT_EXPR, new_ops[1], one_cst); in vect_recog_average_pattern()
2090 g = gimple_build_assign (sum_of_shifted, PLUS_EXPR, in vect_recog_average_pattern()
2096 g = gimple_build_assign (unmasked_carry, c, new_ops[0], new_ops[1]); in vect_recog_average_pattern()
2100 g = gimple_build_assign (carry, BIT_AND_EXPR, unmasked_carry, one_cst); in vect_recog_average_pattern()
2103 g = gimple_build_assign (new_var, PLUS_EXPR, sum_of_shifted, carry); in vect_recog_average_pattern()
2190 gimple *pattern_stmt = gimple_build_assign (new_var, code, unprom.op); in vect_recog_cast_forwprop_pattern()
2318 def_stmt = gimple_build_assign (def, NOP_EXPR, oprnd0); in vect_recog_rotate_pattern()
2378 def_stmt = gimple_build_assign (def, NOP_EXPR, oprnd0); in vect_recog_rotate_pattern()
2391 pattern_stmt = gimple_build_assign (var, LROTATE_EXPR, oprnd0, in vect_recog_rotate_pattern()
2431 def_stmt = gimple_build_assign (def, NOP_EXPR, oprnd0); in vect_recog_rotate_pattern()
2455 def_stmt = gimple_build_assign (def, NOP_EXPR, oprnd1); in vect_recog_rotate_pattern()
2476 def_stmt = gimple_build_assign (def2, NEGATE_EXPR, def); in vect_recog_rotate_pattern()
2488 def_stmt = gimple_build_assign (def2, BIT_AND_EXPR, in vect_recog_rotate_pattern()
2501 def_stmt = gimple_build_assign (var1, rhs_code == LROTATE_EXPR in vect_recog_rotate_pattern()
2507 def_stmt = gimple_build_assign (var2, rhs_code == LROTATE_EXPR in vect_recog_rotate_pattern()
2517 pattern_stmt = gimple_build_assign (var, BIT_IOR_EXPR, var1, var2); in vect_recog_rotate_pattern()
2622 def_stmt = gimple_build_assign (def, BIT_AND_EXPR, rhs1, mask); in vect_recog_vector_vector_shift_pattern()
2633 def_stmt = gimple_build_assign (def, NOP_EXPR, oprnd1); in vect_recog_vector_vector_shift_pattern()
2642 pattern_stmt = gimple_build_assign (var, rhs_code, oprnd0, def); in vect_recog_vector_vector_shift_pattern()
2730 = gimple_build_assign (tmp_var, PLUS_EXPR, prev_res, prev_res); in synth_lshift_by_additions()
2773 stmt = gimple_build_assign (tmp_var, code, op1, op2); in apply_binop_and_append_stmt()
2834 stmt = gimple_build_assign (tmp_op, CONVERT_EXPR, op); in vect_synth_mult_by_constant()
2861 stmt = gimple_build_assign (accum_tmp, LSHIFT_EXPR, accumulator, in vect_synth_mult_by_constant()
2868 stmt = gimple_build_assign (accum_tmp, PLUS_EXPR, accumulator, in vect_synth_mult_by_constant()
2879 stmt = gimple_build_assign (accum_tmp, NEGATE_EXPR, tmp_var); in vect_synth_mult_by_constant()
2881 stmt = gimple_build_assign (accum_tmp, MINUS_EXPR, accumulator, in vect_synth_mult_by_constant()
2888 stmt = gimple_build_assign (accum_tmp, PLUS_EXPR, tmp_var, op); in vect_synth_mult_by_constant()
2894 stmt = gimple_build_assign (accum_tmp, MINUS_EXPR, tmp_var, op); in vect_synth_mult_by_constant()
2900 stmt = gimple_build_assign (accum_tmp, PLUS_EXPR, accumulator, in vect_synth_mult_by_constant()
2907 stmt = gimple_build_assign (accum_tmp, MINUS_EXPR, tmp_var, in vect_synth_mult_by_constant()
2923 stmt = gimple_build_assign (accum_tmp, NEGATE_EXPR, accumulator); in vect_synth_mult_by_constant()
2931 stmt = gimple_build_assign (accum_tmp, PLUS_EXPR, accumulator, op); in vect_synth_mult_by_constant()
2940 stmt = gimple_build_assign (accum_tmp, CONVERT_EXPR, accumulator); in vect_synth_mult_by_constant()
3131 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in vect_recog_divmod_pattern()
3135 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in vect_recog_divmod_pattern()
3154 = gimple_build_assign (var, COND_EXPR, cond, in vect_recog_divmod_pattern()
3161 = gimple_build_assign (var, PLUS_EXPR, oprnd0, in vect_recog_divmod_pattern()
3167 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in vect_recog_divmod_pattern()
3176 def_stmt = gimple_build_assign (signmask, COND_EXPR, cond, in vect_recog_divmod_pattern()
3191 def_stmt = gimple_build_assign (var, COND_EXPR, cond, in vect_recog_divmod_pattern()
3196 def_stmt = gimple_build_assign (var, RSHIFT_EXPR, in vect_recog_divmod_pattern()
3202 = gimple_build_assign (signmask, NOP_EXPR, var); in vect_recog_divmod_pattern()
3206 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in vect_recog_divmod_pattern()
3210 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in vect_recog_divmod_pattern()
3217 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in vect_recog_divmod_pattern()
3271 def_stmt = gimple_build_assign (t1, MULT_HIGHPART_EXPR, oprnd0, in vect_recog_divmod_pattern()
3277 = gimple_build_assign (t2, MINUS_EXPR, oprnd0, t1); in vect_recog_divmod_pattern()
3282 = gimple_build_assign (t3, RSHIFT_EXPR, t2, integer_one_node); in vect_recog_divmod_pattern()
3287 = gimple_build_assign (t4, PLUS_EXPR, t1, t3); in vect_recog_divmod_pattern()
3295 = gimple_build_assign (q, RSHIFT_EXPR, t4, in vect_recog_divmod_pattern()
3316 = gimple_build_assign (t1, RSHIFT_EXPR, oprnd0, in vect_recog_divmod_pattern()
3324 def_stmt = gimple_build_assign (t2, MULT_HIGHPART_EXPR, t1, in vect_recog_divmod_pattern()
3333 = gimple_build_assign (q, RSHIFT_EXPR, t2, in vect_recog_divmod_pattern()
3384 def_stmt = gimple_build_assign (t1, MULT_HIGHPART_EXPR, oprnd0, in vect_recog_divmod_pattern()
3392 def_stmt = gimple_build_assign (t2, PLUS_EXPR, t1, oprnd0); in vect_recog_divmod_pattern()
3402 def_stmt = gimple_build_assign (t3, RSHIFT_EXPR, t2, in vect_recog_divmod_pattern()
3434 def_stmt = gimple_build_assign (t4, INTEGER_CST, in vect_recog_divmod_pattern()
3437 def_stmt = gimple_build_assign (t4, RSHIFT_EXPR, oprnd0, in vect_recog_divmod_pattern()
3443 pattern_stmt = gimple_build_assign (q, MINUS_EXPR, d < 0 ? t4 : t3, in vect_recog_divmod_pattern()
3458 def_stmt = gimple_build_assign (t1, MULT_EXPR, q, oprnd1); in vect_recog_divmod_pattern()
3462 pattern_stmt = gimple_build_assign (r, MINUS_EXPR, oprnd0, t1); in vect_recog_divmod_pattern()
3605 def_stmt = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in vect_recog_mixed_size_cond_pattern()
3609 pattern_stmt = gimple_build_assign (vect_recog_temp_ssa_var (type, NULL), in vect_recog_mixed_size_cond_pattern()
3728 gimple *cast_stmt = gimple_build_assign (vect_recog_temp_ssa_var (type, NULL), in adjust_bool_pattern_cast()
3764 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in adjust_bool_pattern()
3772 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in adjust_bool_pattern()
3874 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in adjust_bool_pattern()
3899 = gimple_build_assign (vect_recog_temp_ssa_var (itype, NULL), in adjust_bool_pattern()
4053 pattern_stmt = gimple_build_assign (lhs, SSA_NAME, rhs); in vect_recog_bool_pattern()
4056 = gimple_build_assign (lhs, NOP_EXPR, rhs); in vect_recog_bool_pattern()
4077 pattern_stmt = gimple_build_assign (tmp, COND_EXPR, var, cst1, cst0); in vect_recog_bool_pattern()
4085 pattern_stmt = gimple_build_assign (lhs, CONVERT_EXPR, tmp); in vect_recog_bool_pattern()
4121 = gimple_build_assign (lhs, COND_EXPR, in vect_recog_bool_pattern()
4159 pattern_stmt = gimple_build_assign (rhs, COND_EXPR, var, cst1, cst0); in vect_recog_bool_pattern()
4167 gimple *cast_stmt = gimple_build_assign (rhs2, NOP_EXPR, rhs); in vect_recog_bool_pattern()
4171 pattern_stmt = gimple_build_assign (lhs, SSA_NAME, rhs); in vect_recog_bool_pattern()
4199 stmt = gimple_build_assign (tmp, CONVERT_EXPR, mask); in build_mask_conversion()
4393 pattern_stmt = gimple_build_assign (tmp, rhs1); in vect_recog_mask_conversion_pattern()
4406 pattern_stmt = gimple_build_assign (lhs, COND_EXPR, tmp, in vect_recog_mask_conversion_pattern()
4452 pattern_stmt = gimple_build_assign (lhs, rhs_code, rhs1, rhs2); in vect_recog_mask_conversion_pattern()
4519 gassign *conversion = gimple_build_assign (new_value, CONVERT_EXPR, value); in vect_add_conversion_to_pattern()