Lines Matching refs:bitsize

371   poly_int64 bitsize, bitpos, bytepos;  in find_bswap_or_nop_load()  local
383 base_addr = get_inner_reference (ref, &bitsize, &bitpos, &offset, &mode, in find_bswap_or_nop_load()
422 if (!multiple_p (bitsize, BITS_PER_UNIT)) in find_bswap_or_nop_load()
601 unsigned HOST_WIDE_INT bitsize = tree_to_uhwi (TREE_OPERAND (rhs1, 1)); in find_bswap_or_nop_1() local
604 && bitsize % BITS_PER_UNIT == 0 in find_bswap_or_nop_1()
609 bitpos = TYPE_PRECISION (n->type) - bitpos - bitsize; in find_bswap_or_nop_1()
618 for (unsigned i = 0; i < bitsize / BITS_PER_UNIT; in find_bswap_or_nop_1()
1342 poly_uint64 bitsize; variable
1352 : val (NULL_TREE), base_addr (NULL_TREE), bitsize (0), bitpos (0), in store_operand_info()
1364 unsigned HOST_WIDE_INT bitsize;
1413 : bitsize (bs), bitpos (bp), bitregion_start (brs), bitregion_end (bre), in store_immediate_info()
1809 width = info->bitsize; in merged_store_group()
1953 width = info->bitpos + info->bitsize - start; in merge_into()
1965 if (info->bitpos + info->bitsize > start + width) in merge_overlapping()
1966 width = info->bitpos + info->bitsize - start; in merge_overlapping()
2011 ret = encode_tree_to_bitpos (cst, val, info->bitsize, in apply_stores()
2018 info->bitsize); in apply_stores()
2020 clear_bit_region (m, pos_in_buffer % BITS_PER_UNIT, info->bitsize); in apply_stores()
2028 " at position %d\n", info->bitsize, pos_in_buffer); in apply_stores()
2438 && info->bitpos + info->bitsize > start) in check_no_overlap()
2465 unsigned HOST_WIDE_INT width = m_store_info[first]->bitsize; in try_coalesce_bswap()
2474 width += m_store_info[i]->bitsize; in try_coalesce_bswap()
2542 ? try_size - info->bitsize - bitpos in try_coalesce_bswap()
2581 end = MAX (end, info->bitpos + info->bitsize); in try_coalesce_bswap()
2715 + m_store_info[first_earlier]->bitsize in coalesce_immediate_stores()
2786 info->bitpos + info->bitsize); in coalesce_immediate_stores()
2847 info2->bitpos + info2->bitsize); in coalesce_immediate_stores()
2957 info->bitpos + info->bitsize), in coalesce_immediate_stores()
3012 i, info->bitsize, info->bitpos); in coalesce_immediate_stores()
3139 unsigned HOST_WIDE_INT bitsize) in find_constituent_stores() argument
3145 unsigned HOST_WIDE_INT end = bitpos + bitsize; in find_constituent_stores()
3149 unsigned HOST_WIDE_INT stmt_end = stmt_start + info->bitsize; in find_constituent_stores()
3386 = new split_store (bytepos, gstore->bitsize, align_base); in split_group()
3449 = ROUND_UP (info->bitpos + info->bitsize, BITS_PER_UNIT); in split_group()
3542 && info->bitpos + info->bitsize <= try_bitpos + try_size in split_group()
3546 && (info->bitpos + info->bitsize in split_group()
3634 && TYPE_PRECISION (TREE_TYPE (lhs)) < info->bitsize) in invert_op()
3657 unsigned HOST_WIDE_INT bitsize = info->bitsize; in invert_op() local
3658 unsigned HOST_WIDE_INT prec = bitsize; in invert_op()
3664 && TYPE_PRECISION (TREE_TYPE (lhs)) < bitsize) in invert_op()
3669 gcc_assert (info->bitpos + bitsize > try_bitpos); in invert_op()
3676 bitsize -= try_bitpos - info->bitpos; in invert_op()
3677 if (BYTES_BIG_ENDIAN && prec > bitsize) in invert_op()
3678 prec = bitsize; in invert_op()
3682 if (prec < bitsize) in invert_op()
3688 pos_in_buffer += bitsize - prec; in invert_op()
3692 bitsize = prec; in invert_op()
3694 if (pos_in_buffer + bitsize > split_store->size) in invert_op()
3695 bitsize = split_store->size - pos_in_buffer; in invert_op()
3699 - (pos_in_buffer % BITS_PER_UNIT)), bitsize); in invert_op()
3701 clear_bit_region (p, pos_in_buffer % BITS_PER_UNIT, bitsize); in invert_op()
3744 && group->width == store->bitsize in output_merged_store()
3799 && group->width == group->stores[0]->bitsize in output_merged_store()
4229 && info->bitpos + info->bitsize > try_bitpos) in output_merged_store()
4235 = (try_bitpos + try_size) - (info->bitpos + info->bitsize); in output_merged_store()
4237 if (TYPE_PRECISION (TREE_TYPE (tem)) <= info->bitsize) in output_merged_store()
4240 = build_nonstandard_integer_type (info->bitsize, in output_merged_store()
4247 = (HOST_WIDE_INT_1U << info->bitsize) - 1; in output_merged_store()
4575 poly_int64 bitsize, bitpos; in mem_valid_for_store_merging() local
4580 tree base_addr = get_inner_reference (mem, &bitsize, &bitpos, &offset, &mode, in mem_valid_for_store_merging()
4582 *pbitsize = bitsize; in mem_valid_for_store_merging()
4583 if (known_le (bitsize, 0)) in mem_valid_for_store_merging()
4646 bitregion_end = round_up_to_byte_boundary (bitpos + bitsize); in mem_valid_for_store_merging()
4649 *pbitsize = bitsize; in mem_valid_for_store_merging()
4662 poly_uint64 bitsize, poly_uint64 bitpos, in handled_load() argument
4671 && handled_load (SSA_NAME_DEF_STMT (rhs1), op, bitsize, bitpos, in handled_load()
4691 = mem_valid_for_store_merging (mem, &op->bitsize, &op->bitpos, in handled_load()
4695 && known_eq (op->bitsize, bitsize) in handled_load()
4733 poly_uint64 bitsize, bitpos; in process_store() local
4736 = mem_valid_for_store_merging (lhs, &bitsize, &bitpos, in process_store()
4738 if (known_eq (bitsize, 0U)) in process_store()
4742 || (maybe_gt (bitsize, in process_store()
4766 else if (handled_load (def_stmt, &ops[0], bitsize, bitpos, in process_store()
4794 || !handled_load (def_stmt1, &ops[0], bitsize, bitpos, in process_store()
4806 else if (!handled_load (def_stmt2, &ops[1], bitsize, bitpos, in process_store()
4818 if (bitsize.is_constant (&const_bitsize) in process_store()
4850 && bitsize.is_constant (&const_bitsize) in process_store()
4877 || !bitsize.is_constant (&const_bitsize) in process_store()