Lines Matching refs:mode

93 mask_rtx (scalar_int_mode mode, int bitpos, int bitsize, bool complement)  in mask_rtx()  argument
97 GET_MODE_PRECISION (mode)), mode); in mask_rtx()
160 machine_mode mode, int speed) in init_expmed_one_mode() argument
165 mode_bitsize = GET_MODE_UNIT_BITSIZE (mode); in init_expmed_one_mode()
167 PUT_MODE (all->reg, mode); in init_expmed_one_mode()
168 PUT_MODE (all->plus, mode); in init_expmed_one_mode()
169 PUT_MODE (all->neg, mode); in init_expmed_one_mode()
170 PUT_MODE (all->mult, mode); in init_expmed_one_mode()
171 PUT_MODE (all->sdiv, mode); in init_expmed_one_mode()
172 PUT_MODE (all->udiv, mode); in init_expmed_one_mode()
173 PUT_MODE (all->sdiv_32, mode); in init_expmed_one_mode()
174 PUT_MODE (all->smod_32, mode); in init_expmed_one_mode()
175 PUT_MODE (all->wide_trunc, mode); in init_expmed_one_mode()
176 PUT_MODE (all->shift, mode); in init_expmed_one_mode()
177 PUT_MODE (all->shift_mult, mode); in init_expmed_one_mode()
178 PUT_MODE (all->shift_add, mode); in init_expmed_one_mode()
179 PUT_MODE (all->shift_sub0, mode); in init_expmed_one_mode()
180 PUT_MODE (all->shift_sub1, mode); in init_expmed_one_mode()
181 PUT_MODE (all->zext, mode); in init_expmed_one_mode()
182 PUT_MODE (all->trunc, mode); in init_expmed_one_mode()
184 set_add_cost (speed, mode, set_src_cost (all->plus, mode, speed)); in init_expmed_one_mode()
185 set_neg_cost (speed, mode, set_src_cost (all->neg, mode, speed)); in init_expmed_one_mode()
186 set_mul_cost (speed, mode, set_src_cost (all->mult, mode, speed)); in init_expmed_one_mode()
187 set_sdiv_cost (speed, mode, set_src_cost (all->sdiv, mode, speed)); in init_expmed_one_mode()
188 set_udiv_cost (speed, mode, set_src_cost (all->udiv, mode, speed)); in init_expmed_one_mode()
190 set_sdiv_pow2_cheap (speed, mode, (set_src_cost (all->sdiv_32, mode, speed) in init_expmed_one_mode()
191 <= 2 * add_cost (speed, mode))); in init_expmed_one_mode()
192 set_smod_pow2_cheap (speed, mode, (set_src_cost (all->smod_32, mode, speed) in init_expmed_one_mode()
193 <= 4 * add_cost (speed, mode))); in init_expmed_one_mode()
195 set_shift_cost (speed, mode, 0, 0); in init_expmed_one_mode()
197 int cost = add_cost (speed, mode); in init_expmed_one_mode()
198 set_shiftadd_cost (speed, mode, 0, cost); in init_expmed_one_mode()
199 set_shiftsub0_cost (speed, mode, 0, cost); in init_expmed_one_mode()
200 set_shiftsub1_cost (speed, mode, 0, cost); in init_expmed_one_mode()
209 set_shift_cost (speed, mode, m, set_src_cost (all->shift, mode, speed)); in init_expmed_one_mode()
210 set_shiftadd_cost (speed, mode, m, set_src_cost (all->shift_add, mode, in init_expmed_one_mode()
212 set_shiftsub0_cost (speed, mode, m, set_src_cost (all->shift_sub0, mode, in init_expmed_one_mode()
214 set_shiftsub1_cost (speed, mode, m, set_src_cost (all->shift_sub1, mode, in init_expmed_one_mode()
219 if (is_a <scalar_int_mode> (mode, &int_mode_to)) in init_expmed_one_mode()
249 machine_mode mode = QImode; in init_expmed() local
260 all.reg = gen_raw_REG (mode, LAST_VIRTUAL_REGISTER + 1); in init_expmed()
261 all.plus = gen_rtx_PLUS (mode, all.reg, all.reg); in init_expmed()
262 all.neg = gen_rtx_NEG (mode, all.reg); in init_expmed()
263 all.mult = gen_rtx_MULT (mode, all.reg, all.reg); in init_expmed()
264 all.sdiv = gen_rtx_DIV (mode, all.reg, all.reg); in init_expmed()
265 all.udiv = gen_rtx_UDIV (mode, all.reg, all.reg); in init_expmed()
266 all.sdiv_32 = gen_rtx_DIV (mode, all.reg, all.pow2[5]); in init_expmed()
267 all.smod_32 = gen_rtx_MOD (mode, all.reg, all.pow2[5]); in init_expmed()
268 all.zext = gen_rtx_ZERO_EXTEND (mode, all.reg); in init_expmed()
269 all.wide_mult = gen_rtx_MULT (mode, all.zext, all.zext); in init_expmed()
270 all.wide_lshr = gen_rtx_LSHIFTRT (mode, all.wide_mult, all.reg); in init_expmed()
271 all.wide_trunc = gen_rtx_TRUNCATE (mode, all.wide_lshr); in init_expmed()
272 all.shift = gen_rtx_ASHIFT (mode, all.reg, all.reg); in init_expmed()
273 all.shift_mult = gen_rtx_MULT (mode, all.reg, all.reg); in init_expmed()
274 all.shift_add = gen_rtx_PLUS (mode, all.shift_mult, all.reg); in init_expmed()
275 all.shift_sub0 = gen_rtx_MINUS (mode, all.shift_mult, all.reg); in init_expmed()
276 all.shift_sub1 = gen_rtx_MINUS (mode, all.reg, all.shift_mult); in init_expmed()
277 all.trunc = gen_rtx_TRUNCATE (mode, all.reg); in init_expmed()
282 set_zero_cost (speed, set_src_cost (const0_rtx, mode, speed)); in init_expmed()
284 for (mode = MIN_MODE_INT; mode <= MAX_MODE_INT; in init_expmed()
285 mode = (machine_mode)(mode + 1)) in init_expmed()
286 init_expmed_one_mode (&all, mode, speed); in init_expmed()
289 for (mode = MIN_MODE_PARTIAL_INT; mode <= MAX_MODE_PARTIAL_INT; in init_expmed()
290 mode = (machine_mode)(mode + 1)) in init_expmed()
291 init_expmed_one_mode (&all, mode, speed); in init_expmed()
294 for (mode = MIN_MODE_VECTOR_INT; mode <= MAX_MODE_VECTOR_INT; in init_expmed()
295 mode = (machine_mode)(mode + 1)) in init_expmed()
296 init_expmed_one_mode (&all, mode, speed); in init_expmed()
333 negate_rtx (machine_mode mode, rtx x) in negate_rtx() argument
335 rtx result = simplify_unary_operation (NEG, mode, x, mode); in negate_rtx()
338 result = expand_unop (mode, neg_optab, x, NULL_RTX, 0); in negate_rtx()
382 flip_storage_order (machine_mode mode, rtx x) in flip_storage_order() argument
387 if (mode == QImode) in flip_storage_order()
390 if (COMPLEX_MODE_P (mode)) in flip_storage_order()
395 real = flip_storage_order (GET_MODE_INNER (mode), real); in flip_storage_order()
396 imag = flip_storage_order (GET_MODE_INNER (mode), imag); in flip_storage_order()
398 return gen_rtx_CONCAT (mode, real, imag); in flip_storage_order()
404 if (!is_a <scalar_int_mode> (mode, &int_mode)) in flip_storage_order()
406 if (FLOAT_MODE_P (mode) in flip_storage_order()
410 if (!int_mode_for_size (GET_MODE_PRECISION (mode), 0).exists (&int_mode)) in flip_storage_order()
412 sorry ("reverse storage order for %smode", GET_MODE_NAME (mode)); in flip_storage_order()
422 if (int_mode != mode) in flip_storage_order()
423 result = gen_lowpart (mode, result); in flip_storage_order()
435 narrow_bit_field_mem (rtx mem, opt_scalar_int_mode mode, in narrow_bit_field_mem() argument
441 if (mode.exists (&imode)) in narrow_bit_field_mem()
580 machine_mode mode, poly_uint64 *bytenum) in simple_mem_bitfield_p() argument
584 && known_eq (bitsize, GET_MODE_BITSIZE (mode)) in simple_mem_bitfield_p()
585 && (!targetm.slow_unaligned_access (mode, MEM_ALIGN (op0)) in simple_mem_bitfield_p()
586 || (multiple_p (bitnum, GET_MODE_ALIGNMENT (mode)) in simple_mem_bitfield_p()
587 && MEM_ALIGN (op0) >= GET_MODE_ALIGNMENT (mode)))); in simple_mem_bitfield_p()
1215 store_fixed_bit_field_1 (rtx op0, scalar_int_mode mode, in store_fixed_bit_field_1() argument
1231 bitnum = GET_MODE_BITSIZE (mode) - bitsize - bitnum; in store_fixed_bit_field_1()
1254 value = lshift_value (mode, v, bitnum); in store_fixed_bit_field_1()
1259 && bitnum + bitsize != GET_MODE_BITSIZE (mode)); in store_fixed_bit_field_1()
1261 if (value_mode != mode) in store_fixed_bit_field_1()
1262 value = convert_to_mode (mode, value, 1); in store_fixed_bit_field_1()
1265 value = expand_binop (mode, and_optab, value, in store_fixed_bit_field_1()
1266 mask_rtx (mode, 0, bitsize, 0), in store_fixed_bit_field_1()
1269 value = expand_shift (LSHIFT_EXPR, mode, value, in store_fixed_bit_field_1()
1274 value = flip_storage_order (mode, value); in store_fixed_bit_field_1()
1281 temp = force_reg (mode, op0); in store_fixed_bit_field_1()
1285 rtx mask = mask_rtx (mode, bitnum, bitsize, 1); in store_fixed_bit_field_1()
1287 mask = flip_storage_order (mode, mask); in store_fixed_bit_field_1()
1288 temp = expand_binop (mode, and_optab, temp, mask, in store_fixed_bit_field_1()
1290 temp = force_reg (mode, temp); in store_fixed_bit_field_1()
1297 temp = expand_binop (mode, ior_optab, temp, value, in store_fixed_bit_field_1()
1299 temp = force_reg (mode, temp); in store_fixed_bit_field_1()
1468 convert_extracted_bit_field (rtx x, machine_mode mode, in convert_extracted_bit_field() argument
1471 if (GET_MODE (x) == tmode || GET_MODE (x) == mode) in convert_extracted_bit_field()
1500 machine_mode mode, machine_mode tmode) in extract_bit_field_using_extv() argument
1568 return convert_extracted_bit_field (target, mode, tmode, unsignedp); in extract_bit_field_using_extv()
1578 extract_bit_field_as_subreg (machine_mode mode, rtx op0, in extract_bit_field_as_subreg() argument
1583 && known_eq (bitsize, GET_MODE_BITSIZE (mode)) in extract_bit_field_as_subreg()
1585 && TRULY_NOOP_TRUNCATION_MODES_P (mode, GET_MODE (op0))) in extract_bit_field_as_subreg()
1586 return simplify_gen_subreg (mode, op0, GET_MODE (op0), bytenum); in extract_bit_field_as_subreg()
1597 int unsignedp, rtx target, machine_mode mode, in extract_bit_field_1() argument
1605 tmode = mode; in extract_bit_field_1()
1620 && mode == GET_MODE (op0) in extract_bit_field_1()
1625 op0 = flip_storage_order (mode, op0); in extract_bit_field_1()
1675 if (GET_MODE (target) != mode) in extract_bit_field_1()
1736 if (GET_MODE (target) != mode) in extract_bit_field_1()
1744 if (VECTOR_MODE_P (mode)) in extract_bit_field_1()
1746 rtx sub = extract_bit_field_as_subreg (mode, op0, bitsize, bitnum); in extract_bit_field_1()
1787 mode1 = mode; in extract_bit_field_1()
1797 return convert_extracted_bit_field (sub, mode, tmode, unsignedp); in extract_bit_field_1()
1808 return convert_extracted_bit_field (op0, mode, tmode, unsignedp); in extract_bit_field_1()
1829 target, mode, tmode, reverse, fallback_p); in extract_bit_field_1()
1841 rtx target, machine_mode mode, machine_mode tmode, in extract_integral_bit_field() argument
1859 target = gen_reg_rtx (mode); in extract_integral_bit_field()
1864 target = gen_reg_rtx (mode); in extract_integral_bit_field()
1893 mode, word_mode, reverse, fallback_p, NULL); in extract_integral_bit_field()
1926 target = expand_shift (LSHIFT_EXPR, mode, target, in extract_integral_bit_field()
1927 GET_MODE_BITSIZE (mode) - bitsize, NULL_RTX, 0); in extract_integral_bit_field()
1928 return expand_shift (RSHIFT_EXPR, mode, target, in extract_integral_bit_field()
1929 GET_MODE_BITSIZE (mode) - bitsize, NULL_RTX, 0); in extract_integral_bit_field()
1942 return convert_extracted_bit_field (target, mode, tmode, unsignedp); in extract_integral_bit_field()
1965 unsignedp, target, mode, in extract_integral_bit_field()
1980 unsignedp, target, mode, in extract_integral_bit_field()
1998 mode, tmode, reverse, false, NULL); in extract_integral_bit_field()
2014 int_mode = int_mode_for_mode (mode).require (); in extract_integral_bit_field()
2024 target = convert_extracted_bit_field (target, mode, tmode, unsignedp); in extract_integral_bit_field()
2028 target = convert_extracted_bit_field (target, mode, tmode, unsignedp); in extract_integral_bit_field()
2053 int unsignedp, rtx target, machine_mode mode, in extract_bit_field() argument
2085 return convert_extracted_bit_field (result, mode, tmode, unsignedp); in extract_bit_field()
2093 target, mode, tmode, reverse, true, alt_rtl); in extract_bit_field()
2097 target, mode, tmode, reverse, true, alt_rtl); in extract_bit_field()
2118 scalar_int_mode mode; in extract_fixed_bit_field() local
2122 BITS_PER_WORD, MEM_VOLATILE_P (op0), &mode)) in extract_fixed_bit_field()
2128 op0 = narrow_bit_field_mem (op0, mode, bitsize, bitnum, &bitnum); in extract_fixed_bit_field()
2131 mode = op0_mode.require (); in extract_fixed_bit_field()
2133 return extract_fixed_bit_field_1 (tmode, op0, mode, bitsize, bitnum, in extract_fixed_bit_field()
2142 extract_fixed_bit_field_1 (machine_mode tmode, rtx op0, scalar_int_mode mode, in extract_fixed_bit_field_1() argument
2154 bitnum = GET_MODE_BITSIZE (mode) - bitsize - bitnum; in extract_fixed_bit_field_1()
2159 op0 = flip_storage_order (mode, op0); in extract_fixed_bit_field_1()
2169 if (tmode != mode) in extract_fixed_bit_field_1()
2171 op0 = expand_shift (RSHIFT_EXPR, mode, op0, bitnum, subtarget, 1); in extract_fixed_bit_field_1()
2177 if (mode != new_mode) in extract_fixed_bit_field_1()
2183 if (GET_MODE_BITSIZE (mode) != bitnum + bitsize) in extract_fixed_bit_field_1()
2192 op0 = force_reg (mode, op0); in extract_fixed_bit_field_1()
2201 mode = mode_iter.require (); in extract_fixed_bit_field_1()
2202 op0 = convert_to_mode (mode, op0, 0); in extract_fixed_bit_field_1()
2204 if (mode != tmode) in extract_fixed_bit_field_1()
2207 if (GET_MODE_BITSIZE (mode) != (bitsize + bitnum)) in extract_fixed_bit_field_1()
2209 int amount = GET_MODE_BITSIZE (mode) - (bitsize + bitnum); in extract_fixed_bit_field_1()
2212 op0 = expand_shift (LSHIFT_EXPR, mode, op0, amount, subtarget, 1); in extract_fixed_bit_field_1()
2215 return expand_shift (RSHIFT_EXPR, mode, op0, in extract_fixed_bit_field_1()
2216 GET_MODE_BITSIZE (mode) - bitsize, target, 0); in extract_fixed_bit_field_1()
2223 lshift_value (machine_mode mode, unsigned HOST_WIDE_INT value, in lshift_value() argument
2226 return immed_wide_int_const (wi::lshift (value, bitpos), mode); in lshift_value()
2349 extract_low_bits (machine_mode mode, machine_mode src_mode, rtx src) in extract_low_bits() argument
2353 if (mode == src_mode) in extract_low_bits()
2361 poly_uint64 byte = subreg_lowpart_offset (mode, src_mode); in extract_low_bits()
2362 rtx ret = simplify_subreg (mode, src, src_mode, byte); in extract_low_bits()
2367 || !validate_subreg (mode, src_mode, src, byte)) in extract_low_bits()
2371 return gen_rtx_SUBREG (mode, src, byte); in extract_low_bits()
2374 if (GET_MODE_CLASS (mode) == MODE_CC || GET_MODE_CLASS (src_mode) == MODE_CC) in extract_low_bits()
2377 if (known_eq (GET_MODE_BITSIZE (mode), GET_MODE_BITSIZE (src_mode)) in extract_low_bits()
2378 && targetm.modes_tieable_p (mode, src_mode)) in extract_low_bits()
2380 rtx x = gen_lowpart_common (mode, src); in extract_low_bits()
2386 || !int_mode_for_mode (mode).exists (&int_mode)) in extract_low_bits()
2391 if (!targetm.modes_tieable_p (int_mode, mode)) in extract_low_bits()
2396 src = gen_lowpart (mode, src); in extract_low_bits()
2434 expand_shift_1 (enum tree_code code, machine_mode mode, rtx shifted,
2446 scalar_mode scalar_mode = GET_MODE_INNER (mode);
2455 if (VECTOR_MODE_P (mode) && VECTOR_MODE_P (op1_mode))
2473 op1 = gen_int_shift_amount (mode,
2492 op1 = gen_int_shift_amount (mode, (GET_MODE_BITSIZE (scalar_mode)
2506 && optab_handler (bswap_optab, mode) != CODE_FOR_nothing)
2507 return expand_unop (mode, bswap_optab, shifted, NULL_RTX, unsignedp);
2519 && (shift_cost (speed, mode, INTVAL (op1))
2520 > INTVAL (op1) * add_cost (speed, mode))
2521 && shift_cost (speed, mode, INTVAL (op1)) != MAX_COST)
2526 temp = force_reg (mode, shifted);
2527 shifted = expand_binop (mode, add_optab, temp, temp, NULL_RTX,
2573 (mode, GET_MODE_BITSIZE (scalar_mode) - INTVAL (op1));
2585 shifted = force_reg (mode, shifted);
2588 mode, shifted, new_amount, 0, 1);
2590 mode, shifted, other_amount,
2592 return expand_binop (mode, ior_optab, temp, temp1, target,
2596 temp = expand_binop (mode,
2601 temp = expand_binop (mode,
2620 temp = expand_binop (mode,
2643 expand_shift (enum tree_code code, machine_mode mode, rtx shifted, in expand_shift() argument
2646 return expand_shift_1 (code, mode, shifted, in expand_shift()
2647 gen_int_shift_amount (mode, amount), in expand_shift()
2654 maybe_expand_shift (enum tree_code code, machine_mode mode, rtx shifted, in maybe_expand_shift() argument
2657 return expand_shift_1 (code, mode, in maybe_expand_shift()
2669 expand_variable_shift (enum tree_code code, machine_mode mode, rtx shifted, in expand_variable_shift() argument
2672 return expand_shift_1 (code, mode, in expand_variable_shift()
2678 const struct mult_cost *, machine_mode mode);
2694 const struct mult_cost *cost_limit, machine_mode mode) in synth_mult() argument
2720 imode = as_a <scalar_int_mode> (GET_MODE_INNER (mode)); in synth_mult()
2760 hash_index = (t ^ (unsigned int) mode ^ (speed * 256)) % NUM_ALG_HASH_ENTRIES; in synth_mult()
2765 && entry_ptr->mode == mode in synth_mult()
2836 op_cost = m * add_cost (speed, mode); in synth_mult()
2837 if (shift_cost (speed, mode, m) < op_cost) in synth_mult()
2838 op_cost = shift_cost (speed, mode, m); in synth_mult()
2841 synth_mult (alg_in, q, &new_limit, mode); in synth_mult()
2866 op_cost = m * add_cost (speed, mode); in synth_mult()
2867 if (shift_cost (speed, mode, m) < op_cost) in synth_mult()
2868 op_cost = shift_cost (speed, mode, m); in synth_mult()
2871 synth_mult (alg_in, q, &new_limit, mode); in synth_mult()
2909 op_cost = add_cost (speed, mode); in synth_mult()
2912 synth_mult (alg_in, t + 1, &new_limit, mode); in synth_mult()
2928 op_cost = add_cost (speed, mode); in synth_mult()
2931 synth_mult (alg_in, t - 1, &new_limit, mode); in synth_mult()
2949 op_cost = add_cost (speed, mode) + shift_cost (speed, mode, m); in synth_mult()
2956 if (shiftsub1_cost (speed, mode, m) <= op_cost) in synth_mult()
2958 op_cost = shiftsub1_cost (speed, mode, m); in synth_mult()
2962 op_latency = add_cost (speed, mode); in synth_mult()
2967 &new_limit, mode); in synth_mult()
3003 op_cost = add_cost (speed, mode) + shift_cost (speed, mode, m); in synth_mult()
3004 if (shiftadd_cost (speed, mode, m) <= op_cost) in synth_mult()
3005 op_cost = shiftadd_cost (speed, mode, m); in synth_mult()
3012 synth_mult (alg_in, t / d, &new_limit, mode); in synth_mult()
3033 op_cost = add_cost (speed, mode) + shift_cost (speed, mode, m); in synth_mult()
3034 if (shiftsub0_cost (speed, mode, m) <= op_cost) in synth_mult()
3035 op_cost = shiftsub0_cost (speed, mode, m); in synth_mult()
3041 synth_mult (alg_in, t / d, &new_limit, mode); in synth_mult()
3069 op_cost = shiftadd_cost (speed, mode, m); in synth_mult()
3072 synth_mult (alg_in, (t - 1) >> m, &new_limit, mode); in synth_mult()
3092 op_cost = shiftsub0_cost (speed, mode, m); in synth_mult()
3095 synth_mult (alg_in, (t + 1) >> m, &new_limit, mode); in synth_mult()
3121 entry_ptr->mode = mode; in synth_mult()
3132 entry_ptr->mode = mode; in synth_mult()
3166 choose_mult_variant (machine_mode mode, HOST_WIDE_INT val, in choose_mult_variant() argument
3182 op_cost = 2 * GET_MODE_UNIT_BITSIZE (mode) * add_cost (speed, mode); in choose_mult_variant()
3189 synth_mult (alg, val, &limit, mode); in choose_mult_variant()
3193 if (HOST_BITS_PER_INT >= GET_MODE_UNIT_BITSIZE (mode)) in choose_mult_variant()
3195 op_cost = neg_cost (speed, mode); in choose_mult_variant()
3207 synth_mult (&alg2, -val, &limit, mode); in choose_mult_variant()
3215 op_cost = add_cost (speed, mode); in choose_mult_variant()
3227 synth_mult (&alg2, val - 1, &limit, mode); in choose_mult_variant()
3242 expand_mult_const (machine_mode mode, rtx op0, HOST_WIDE_INT val, in expand_mult_const() argument
3254 op0 = force_reg (mode, op0); in expand_mult_const()
3261 accum = copy_to_mode_reg (mode, CONST0_RTX (mode)); in expand_mult_const()
3266 accum = copy_to_mode_reg (mode, op0); in expand_mult_const()
3286 tem = expand_shift (LSHIFT_EXPR, mode, accum, log, NULL_RTX, 0); in expand_mult_const()
3293 tem = expand_shift (LSHIFT_EXPR, mode, op0, log, NULL_RTX, 0); in expand_mult_const()
3294 accum = force_operand (gen_rtx_PLUS (mode, accum, tem), in expand_mult_const()
3300 tem = expand_shift (LSHIFT_EXPR, mode, op0, log, NULL_RTX, 0); in expand_mult_const()
3301 accum = force_operand (gen_rtx_MINUS (mode, accum, tem), in expand_mult_const()
3307 accum = expand_shift (LSHIFT_EXPR, mode, accum, in expand_mult_const()
3309 accum = force_operand (gen_rtx_PLUS (mode, accum, op0), in expand_mult_const()
3315 accum = expand_shift (LSHIFT_EXPR, mode, accum, in expand_mult_const()
3317 accum = force_operand (gen_rtx_MINUS (mode, accum, op0), in expand_mult_const()
3323 tem = expand_shift (LSHIFT_EXPR, mode, accum, log, NULL_RTX, 0); in expand_mult_const()
3324 accum = force_operand (gen_rtx_PLUS (mode, accum, tem), in expand_mult_const()
3330 tem = expand_shift (LSHIFT_EXPR, mode, accum, log, NULL_RTX, 0); in expand_mult_const()
3331 accum = force_operand (gen_rtx_MINUS (mode, tem, accum), in expand_mult_const()
3341 if (SCALAR_INT_MODE_P (mode)) in expand_mult_const()
3346 tem = op0, nmode = mode; in expand_mult_const()
3368 accum = expand_unop (mode, neg_optab, accum, target, 0); in expand_mult_const()
3373 accum = force_operand (gen_rtx_PLUS (mode, accum, op0), target); in expand_mult_const()
3378 nmode = GET_MODE_INNER (mode); in expand_mult_const()
3395 expand_mult (machine_mode mode, rtx op0, rtx op1, rtx target, in expand_mult() argument
3403 bool do_trapv = flag_trapv && SCALAR_INT_MODE_P (mode) && !unsignedp; in expand_mult()
3412 if (INTEGRAL_MODE_P (mode)) in expand_mult()
3419 if (op1 == CONST0_RTX (mode)) in expand_mult()
3421 if (op1 == CONST1_RTX (mode)) in expand_mult()
3423 if (op1 == CONSTM1_RTX (mode)) in expand_mult()
3424 return expand_unop (mode, do_trapv ? negv_optab : neg_optab, in expand_mult()
3433 if (GET_MODE_CLASS (mode) == MODE_VECTOR_INT in expand_mult()
3434 && optab_handler (vashl_optab, mode) == CODE_FOR_nothing in expand_mult()
3435 && optab_handler (ashl_optab, mode) == CODE_FOR_nothing) in expand_mult()
3440 mode_bitsize = GET_MODE_UNIT_BITSIZE (mode); in expand_mult()
3459 int shift = wi::exact_log2 (rtx_mode_t (scalar_op1, mode)); in expand_mult()
3462 return expand_shift (LSHIFT_EXPR, mode, op0, in expand_mult()
3478 return expand_shift (LSHIFT_EXPR, mode, op0, in expand_mult()
3481 fake_reg = gen_raw_REG (mode, LAST_VIRTUAL_REGISTER + 1); in expand_mult()
3493 max_cost = (set_src_cost (gen_rtx_MULT (mode, fake_reg, op1), in expand_mult()
3494 mode, speed) in expand_mult()
3495 - neg_cost (speed, mode)); in expand_mult()
3502 rtx temp = expand_shift (LSHIFT_EXPR, mode, op0, in expand_mult()
3504 return expand_unop (mode, neg_optab, temp, target, 0); in expand_mult()
3507 if (choose_mult_variant (mode, coeff, &algorithm, &variant, in expand_mult()
3510 rtx temp = expand_mult_const (mode, op0, coeff, NULL_RTX, in expand_mult()
3512 return expand_unop (mode, neg_optab, temp, target, 0); in expand_mult()
3519 max_cost = set_src_cost (gen_rtx_MULT (mode, fake_reg, op1), mode, speed); in expand_mult()
3520 if (choose_mult_variant (mode, coeff, &algorithm, &variant, max_cost)) in expand_mult()
3521 return expand_mult_const (mode, op0, coeff, target, in expand_mult()
3531 return expand_binop (mode, add_optab, op0, op0, in expand_mult()
3538 op0 = expand_binop (mode, do_trapv ? smulv_optab : smul_optab, in expand_mult()
3549 mult_by_coeff_cost (HOST_WIDE_INT coeff, machine_mode mode, bool speed) in mult_by_coeff_cost() argument
3555 rtx fake_reg = gen_raw_REG (mode, LAST_VIRTUAL_REGISTER + 1); in mult_by_coeff_cost()
3556 max_cost = set_src_cost (gen_rtx_MULT (mode, fake_reg, fake_reg), in mult_by_coeff_cost()
3557 mode, speed); in mult_by_coeff_cost()
3558 if (choose_mult_variant (mode, coeff, &algorithm, &variant, max_cost)) in mult_by_coeff_cost()
3575 expand_widening_mult (machine_mode mode, rtx op0, rtx op1, rtx target, in expand_widening_mult() argument
3583 && (cop1 = convert_modes (mode, GET_MODE (op0), op1, in expand_widening_mult()
3587 || HWI_COMPUTABLE_MODE_P (mode))) in expand_widening_mult()
3595 return CONST0_RTX (mode); in expand_widening_mult()
3600 op0 = convert_to_mode (mode, op0, this_optab == umul_widen_optab); in expand_widening_mult()
3601 return expand_shift (LSHIFT_EXPR, mode, op0, in expand_widening_mult()
3607 max_cost = mul_widen_cost (speed, mode); in expand_widening_mult()
3608 if (choose_mult_variant (mode, coeff, &algorithm, &variant, in expand_widening_mult()
3611 op0 = convert_to_mode (mode, op0, this_optab == umul_widen_optab); in expand_widening_mult()
3612 return expand_mult_const (mode, op0, coeff, target, in expand_widening_mult()
3616 return expand_binop (mode, this_optab, op0, op1, target, in expand_widening_mult()
3731 expand_mult_highpart_adjust (scalar_int_mode mode, rtx adj_operand, rtx op0, in expand_mult_highpart_adjust() argument
3737 tem = expand_shift (RSHIFT_EXPR, mode, op0, in expand_mult_highpart_adjust()
3738 GET_MODE_BITSIZE (mode) - 1, NULL_RTX, 0); in expand_mult_highpart_adjust()
3739 tem = expand_and (mode, tem, op1, NULL_RTX); in expand_mult_highpart_adjust()
3741 = force_operand (gen_rtx_fmt_ee (adj_code, mode, adj_operand, tem), in expand_mult_highpart_adjust()
3744 tem = expand_shift (RSHIFT_EXPR, mode, op1, in expand_mult_highpart_adjust()
3745 GET_MODE_BITSIZE (mode) - 1, NULL_RTX, 0); in expand_mult_highpart_adjust()
3746 tem = expand_and (mode, tem, op0, NULL_RTX); in expand_mult_highpart_adjust()
3747 target = force_operand (gen_rtx_fmt_ee (adj_code, mode, adj_operand, tem), in expand_mult_highpart_adjust()
3756 extract_high_half (scalar_int_mode mode, rtx op) in extract_high_half() argument
3758 if (mode == word_mode) in extract_high_half()
3759 return gen_highpart (mode, op); in extract_high_half()
3761 scalar_int_mode wider_mode = GET_MODE_WIDER_MODE (mode).require (); in extract_high_half()
3764 GET_MODE_BITSIZE (mode), 0, 1); in extract_high_half()
3765 return convert_modes (mode, wider_mode, op, 0); in extract_high_half()
3772 expmed_mult_highpart_optab (scalar_int_mode mode, rtx op0, rtx op1, in expmed_mult_highpart_optab() argument
3775 rtx narrow_op1 = gen_int_mode (INTVAL (op1), mode); in expmed_mult_highpart_optab()
3781 scalar_int_mode wider_mode = GET_MODE_WIDER_MODE (mode).require (); in expmed_mult_highpart_optab()
3783 size = GET_MODE_BITSIZE (mode); in expmed_mult_highpart_optab()
3787 if (mul_highpart_cost (speed, mode) < max_cost) in expmed_mult_highpart_optab()
3790 tem = expand_binop (mode, moptab, op0, narrow_op1, target, in expmed_mult_highpart_optab()
3799 && (mul_highpart_cost (speed, mode) in expmed_mult_highpart_optab()
3800 + 2 * shift_cost (speed, mode, size-1) in expmed_mult_highpart_optab()
3801 + 4 * add_cost (speed, mode) < max_cost)) in expmed_mult_highpart_optab()
3804 tem = expand_binop (mode, moptab, op0, narrow_op1, target, in expmed_mult_highpart_optab()
3808 return expand_mult_highpart_adjust (mode, tem, op0, narrow_op1, in expmed_mult_highpart_optab()
3814 if (convert_optab_handler (moptab, wider_mode, mode) != CODE_FOR_nothing in expmed_mult_highpart_optab()
3820 return extract_high_half (mode, tem); in expmed_mult_highpart_optab()
3826 && (mul_cost (speed, wider_mode) + shift_cost (speed, mode, size-1) in expmed_mult_highpart_optab()
3837 wop0 = convert_modes (wider_mode, mode, op0, unsignedp); in expmed_mult_highpart_optab()
3838 wop1 = convert_modes (wider_mode, mode, op1, unsignedp); in expmed_mult_highpart_optab()
3847 return extract_high_half (mode, tem); in expmed_mult_highpart_optab()
3853 if (convert_optab_handler (moptab, wider_mode, mode) != CODE_FOR_nothing in expmed_mult_highpart_optab()
3856 + 2 * shift_cost (speed, mode, size-1) in expmed_mult_highpart_optab()
3857 + 4 * add_cost (speed, mode) < max_cost)) in expmed_mult_highpart_optab()
3863 tem = extract_high_half (mode, tem); in expmed_mult_highpart_optab()
3865 return expand_mult_highpart_adjust (mode, tem, op0, narrow_op1, in expmed_mult_highpart_optab()
3885 expmed_mult_highpart (scalar_int_mode mode, rtx op0, rtx op1, in expmed_mult_highpart() argument
3897 gcc_assert (HWI_COMPUTABLE_MODE_P (mode)); in expmed_mult_highpart()
3899 cnst1 = INTVAL (op1) & GET_MODE_MASK (mode); in expmed_mult_highpart()
3905 scalar_int_mode wider_mode = GET_MODE_WIDER_MODE (mode).require (); in expmed_mult_highpart()
3907 return expmed_mult_highpart_optab (mode, op0, op1, target, in expmed_mult_highpart()
3910 extra_cost = shift_cost (speed, mode, GET_MODE_BITSIZE (mode) - 1); in expmed_mult_highpart()
3913 if (!unsignedp && ((cnst1 >> (GET_MODE_BITSIZE (mode) - 1)) & 1)) in expmed_mult_highpart()
3916 extra_cost += add_cost (speed, mode); in expmed_mult_highpart()
3925 tem = expmed_mult_highpart_optab (mode, op0, op1, target, unsignedp, in expmed_mult_highpart()
3932 tem = extract_high_half (mode, tem); in expmed_mult_highpart()
3936 tem = force_operand (gen_rtx_MINUS (mode, tem, op0), tem); in expmed_mult_highpart()
3940 return expmed_mult_highpart_optab (mode, op0, op1, target, in expmed_mult_highpart()
3948 expand_smod_pow2 (scalar_int_mode mode, rtx op0, HOST_WIDE_INT d) in expand_smod_pow2() argument
3953 int prec = GET_MODE_PRECISION (mode); in expand_smod_pow2()
3956 result = gen_reg_rtx (mode); in expand_smod_pow2()
3963 mode, 0, -1); in expand_smod_pow2()
3967 signmask = force_reg (mode, signmask); in expand_smod_pow2()
3968 shift = gen_int_shift_amount (mode, GET_MODE_BITSIZE (mode) - logd); in expand_smod_pow2()
3975 temp = gen_rtx_LSHIFTRT (mode, result, shift); in expand_smod_pow2()
3976 if (optab_handler (lshr_optab, mode) == CODE_FOR_nothing in expand_smod_pow2()
3977 || (set_src_cost (temp, mode, optimize_insn_for_speed_p ()) in expand_smod_pow2()
3980 temp = expand_binop (mode, xor_optab, op0, signmask, in expand_smod_pow2()
3982 temp = expand_binop (mode, sub_optab, temp, signmask, in expand_smod_pow2()
3984 temp = expand_binop (mode, and_optab, temp, in expand_smod_pow2()
3985 gen_int_mode (masklow, mode), in expand_smod_pow2()
3987 temp = expand_binop (mode, xor_optab, temp, signmask, in expand_smod_pow2()
3989 temp = expand_binop (mode, sub_optab, temp, signmask, in expand_smod_pow2()
3994 signmask = expand_binop (mode, lshr_optab, signmask, shift, in expand_smod_pow2()
3996 signmask = force_reg (mode, signmask); in expand_smod_pow2()
3998 temp = expand_binop (mode, add_optab, op0, signmask, in expand_smod_pow2()
4000 temp = expand_binop (mode, and_optab, temp, in expand_smod_pow2()
4001 gen_int_mode (masklow, mode), in expand_smod_pow2()
4003 temp = expand_binop (mode, sub_optab, temp, signmask, in expand_smod_pow2()
4017 temp = expand_binop (mode, and_optab, op0, in expand_smod_pow2()
4018 immed_wide_int_const (mask, mode), in expand_smod_pow2()
4024 do_cmp_and_jump (result, const0_rtx, GE, mode, label); in expand_smod_pow2()
4026 temp = expand_binop (mode, sub_optab, result, const1_rtx, result, in expand_smod_pow2()
4030 temp = expand_binop (mode, ior_optab, temp, in expand_smod_pow2()
4031 immed_wide_int_const (mask, mode), in expand_smod_pow2()
4033 temp = expand_binop (mode, add_optab, temp, const1_rtx, result, in expand_smod_pow2()
4045 expand_sdiv_pow2 (scalar_int_mode mode, rtx op0, HOST_WIDE_INT d) in expand_sdiv_pow2() argument
4057 temp = gen_reg_rtx (mode); in expand_sdiv_pow2()
4058 temp = emit_store_flag (temp, LT, op0, const0_rtx, mode, 0, 1); in expand_sdiv_pow2()
4059 temp = expand_binop (mode, add_optab, temp, op0, NULL_RTX, in expand_sdiv_pow2()
4061 return expand_shift (RSHIFT_EXPR, mode, temp, logd, NULL_RTX, 0); in expand_sdiv_pow2()
4070 temp2 = copy_to_mode_reg (mode, op0); in expand_sdiv_pow2()
4071 temp = expand_binop (mode, add_optab, temp2, gen_int_mode (d - 1, mode), in expand_sdiv_pow2()
4073 temp = force_reg (mode, temp); in expand_sdiv_pow2()
4077 mode, temp, temp2, mode, 0); in expand_sdiv_pow2()
4083 return expand_shift (RSHIFT_EXPR, mode, temp2, logd, NULL_RTX, 0); in expand_sdiv_pow2()
4091 int ushift = GET_MODE_BITSIZE (mode) - logd; in expand_sdiv_pow2()
4093 temp = gen_reg_rtx (mode); in expand_sdiv_pow2()
4094 temp = emit_store_flag (temp, LT, op0, const0_rtx, mode, 0, -1); in expand_sdiv_pow2()
4095 if (GET_MODE_BITSIZE (mode) >= BITS_PER_WORD in expand_sdiv_pow2()
4096 || shift_cost (optimize_insn_for_speed_p (), mode, ushift) in expand_sdiv_pow2()
4098 temp = expand_binop (mode, and_optab, temp, gen_int_mode (d - 1, mode), in expand_sdiv_pow2()
4101 temp = expand_shift (RSHIFT_EXPR, mode, temp, in expand_sdiv_pow2()
4103 temp = expand_binop (mode, add_optab, temp, op0, NULL_RTX, in expand_sdiv_pow2()
4105 return expand_shift (RSHIFT_EXPR, mode, temp, logd, NULL_RTX, 0); in expand_sdiv_pow2()
4109 temp = copy_to_mode_reg (mode, op0); in expand_sdiv_pow2()
4110 do_cmp_and_jump (temp, const0_rtx, GE, mode, label); in expand_sdiv_pow2()
4111 expand_inc (temp, gen_int_mode (d - 1, mode)); in expand_sdiv_pow2()
4113 return expand_shift (RSHIFT_EXPR, mode, temp, logd, NULL_RTX, 0); in expand_sdiv_pow2()
4155 expand_divmod (int rem_flag, enum tree_code code, machine_mode mode, in expand_divmod() argument
4172 wide_int ext_op1 = rtx_mode_t (op1, mode); in expand_divmod()
4219 return expand_unop (mode, flag_trapv && GET_MODE_CLASS (mode) == MODE_INT in expand_divmod()
4262 FOR_EACH_MODE_FROM (compute_mode, mode) in expand_divmod()
4268 FOR_EACH_MODE_FROM (compute_mode, mode) in expand_divmod()
4276 compute_mode = mode; in expand_divmod()
4287 size = GET_MODE_BITSIZE (mode); in expand_divmod()
4304 if (compute_mode != mode) in expand_divmod()
4306 op0 = convert_modes (compute_mode, mode, op0, unsignedp); in expand_divmod()
4307 op1 = convert_modes (compute_mode, mode, op1, unsignedp); in expand_divmod()
4378 return gen_lowpart (mode, remainder); in expand_divmod()
4538 return gen_lowpart (mode, remainder); in expand_divmod()
4686 return gen_lowpart (mode, remainder); in expand_divmod()
4789 return gen_lowpart (mode, rem_flag ? remainder : quotient); in expand_divmod()
4905 return gen_lowpart (mode, rem_flag ? remainder : quotient); in expand_divmod()
5005 return gen_lowpart (mode, rem_flag ? remainder : quotient); in expand_divmod()
5147 return gen_lowpart (mode, rem_flag ? remainder : quotient); in expand_divmod()
5190 return gen_lowpart (mode, remainder); in expand_divmod()
5261 return gen_lowpart (mode, rem_flag ? remainder : quotient); in expand_divmod()
5397 expand_and (machine_mode mode, rtx op0, rtx op1, rtx target) in expand_and() argument
5402 tem = simplify_binary_operation (AND, mode, op0, op1); in expand_and()
5404 tem = expand_binop (mode, and_optab, op0, op1, target, 0, OPTAB_LIB_WIDEN); in expand_and()
5416 machine_mode mode, machine_mode compare_mode, in emit_cstore() argument
5427 x = prepare_operand (icode, x, 2, mode, compare_mode, unsignedp); in emit_cstore()
5428 y = prepare_operand (icode, y, 3, mode, compare_mode, unsignedp); in emit_cstore()
5521 machine_mode mode, int unsignedp, int normalizep, in emit_store_flag_1() argument
5543 if (mode == VOIDmode) in emit_store_flag_1()
5544 mode = GET_MODE (op0); in emit_store_flag_1()
5583 if (is_int_mode (mode, &int_mode) in emit_store_flag_1()
5638 && is_int_mode (mode, &int_mode) in emit_store_flag_1()
5682 mclass = GET_MODE_CLASS (mode); in emit_store_flag_1()
5683 FOR_EACH_MODE_FROM (compare_mode, mode) in emit_store_flag_1()
5690 rtx tem = emit_cstore (target, icode, code, mode, compare_mode, in emit_store_flag_1()
5695 if (GET_MODE_CLASS (mode) == MODE_FLOAT) in emit_store_flag_1()
5697 tem = emit_cstore (target, icode, scode, mode, compare_mode, in emit_store_flag_1()
5716 rtx op1, scalar_int_mode mode, int unsignedp, in emit_store_flag_int() argument
5729 rtx tem = expand_binop (mode, xor_optab, op0, op1, subtarget, 1, in emit_store_flag_int()
5733 tem = expand_binop (mode, sub_optab, op0, op1, subtarget, 1, in emit_store_flag_int()
5737 mode, unsignedp, normalizep); in emit_store_flag_int()
5748 if (can_compare_p (rcode, mode, ccp_store_flag) in emit_store_flag_int()
5749 && ! (optab_handler (cstore_optab, mode) == CODE_FOR_nothing in emit_store_flag_int()
5751 && GET_MODE_SIZE (mode) < UNITS_PER_WORD in emit_store_flag_int()
5759 && rtx_cost (GEN_INT (normalizep), mode, PLUS, 1, in emit_store_flag_int()
5762 rtx tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, in emit_store_flag_int()
5772 && rtx_cost (trueval, mode, XOR, 1, in emit_store_flag_int()
5775 rtx tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, in emit_store_flag_int()
5812 tem = expand_binop (mode, sub_optab, op0, const1_rtx, subtarget, 0, in emit_store_flag_int()
5815 tem = expand_binop (mode, ior_optab, op0, tem, subtarget, 0, in emit_store_flag_int()
5827 tem = maybe_expand_shift (RSHIFT_EXPR, mode, op0, in emit_store_flag_int()
5828 GET_MODE_BITSIZE (mode) - 1, in emit_store_flag_int()
5831 tem = expand_binop (mode, sub_optab, tem, op0, subtarget, 0, in emit_store_flag_int()
5852 if (optab_handler (abs_optab, mode) != CODE_FOR_nothing) in emit_store_flag_int()
5853 tem = expand_unop (mode, abs_optab, op0, subtarget, 1); in emit_store_flag_int()
5854 else if (optab_handler (ffs_optab, mode) != CODE_FOR_nothing) in emit_store_flag_int()
5855 tem = expand_unop (mode, ffs_optab, op0, subtarget, 1); in emit_store_flag_int()
5856 else if (GET_MODE_SIZE (mode) < UNITS_PER_WORD) in emit_store_flag_int()
5858 tem = convert_modes (word_mode, mode, op0, 1); in emit_store_flag_int()
5859 mode = word_mode; in emit_store_flag_int()
5865 tem = expand_binop (mode, sub_optab, tem, const1_rtx, subtarget, in emit_store_flag_int()
5868 tem = expand_unop (mode, neg_optab, tem, subtarget, 0); in emit_store_flag_int()
5884 tem = expand_unop (mode, neg_optab, op0, subtarget, 0); in emit_store_flag_int()
5885 tem = expand_binop (mode, ior_optab, tem, op0, subtarget, 0, in emit_store_flag_int()
5889 tem = expand_unop (mode, one_cmpl_optab, tem, subtarget, 0); in emit_store_flag_int()
5894 tem = maybe_expand_shift (RSHIFT_EXPR, mode, tem, in emit_store_flag_int()
5895 GET_MODE_BITSIZE (mode) - 1, in emit_store_flag_int()
5936 machine_mode mode, int unsignedp, int normalizep) in emit_store_flag() argument
5951 tem = emit_store_flag_1 (target, code, op0, op1, mode, unsignedp, normalizep, in emit_store_flag()
5970 else if (val_signbit_p (mode, STORE_FLAG_VALUE)) in emit_store_flag()
5982 && (target_mode == mode)) ? target : NULL_RTX; in emit_store_flag()
5987 if (GET_MODE_CLASS (mode) == MODE_FLOAT) in emit_store_flag()
5993 if (can_compare_p (rcode, mode, ccp_store_flag) in emit_store_flag()
5995 || (! HONOR_NANS (mode) && (code == LTGT || code == UNEQ)) in emit_store_flag()
5996 || (! HONOR_SNANS (mode) && (code == EQ || code == NE)))) in emit_store_flag()
6003 && rtx_cost (GEN_INT (normalizep), mode, PLUS, 1, in emit_store_flag()
6006 tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, in emit_store_flag()
6014 && rtx_cost (trueval, mode, XOR, 1, in emit_store_flag()
6017 tem = emit_store_flag_1 (subtarget, rcode, op0, op1, mode, 0, in emit_store_flag()
6032 and_them = split_comparison (code, mode, &first_code, &code); in emit_store_flag()
6036 if (!HONOR_NANS (mode)) in emit_store_flag()
6039 return emit_store_flag_1 (target, code, op0, op1, mode, 0, normalizep, in emit_store_flag()
6053 tem = emit_store_flag_1 (subtarget, first_code, op0, op1, mode, 0, in emit_store_flag()
6059 tem = emit_conditional_move (target, code, op0, op1, mode, in emit_store_flag()
6062 tem = emit_conditional_move (target, code, op0, op1, mode, in emit_store_flag()
6073 if (is_int_mode (mode, &int_mode)) in emit_store_flag()
6084 machine_mode mode, int unsignedp, int normalizep) in emit_store_flag_force() argument
6091 tem = emit_store_flag (target, code, op0, op1, mode, unsignedp, normalizep); in emit_store_flag_force()
6103 if (mode == VOIDmode) in emit_store_flag_force()
6104 mode = GET_MODE (op0); in emit_store_flag_force()
6113 && GET_MODE_CLASS (mode) == MODE_INT in emit_store_flag_force()
6119 do_compare_rtx_and_jump (target, const0_rtx, EQ, unsignedp, mode, in emit_store_flag_force()
6134 if (! can_compare_p (code, mode, ccp_jump) in emit_store_flag_force()
6135 && (! FLOAT_MODE_P (mode) in emit_store_flag_force()
6137 || (! HONOR_NANS (mode) && (code == LTGT || code == UNEQ)) in emit_store_flag_force()
6138 || (! HONOR_SNANS (mode) && (code == EQ || code == NE)))) in emit_store_flag_force()
6141 if (FLOAT_MODE_P (mode)) in emit_store_flag_force()
6147 if (can_compare_p (rcode, mode, ccp_jump) in emit_store_flag_force()
6148 || (code == ORDERED && ! can_compare_p (ORDERED, mode, ccp_jump))) in emit_store_flag_force()
6158 do_compare_rtx_and_jump (op0, op1, code, unsignedp, mode, NULL_RTX, NULL, in emit_store_flag_force()
6172 do_cmp_and_jump (rtx arg1, rtx arg2, enum rtx_code op, machine_mode mode, in do_cmp_and_jump() argument
6176 do_compare_rtx_and_jump (arg1, arg2, op, unsignedp, mode, NULL_RTX, in do_cmp_and_jump()