| /src/external/gpl3/gcc/dist/gcc/ |
| gimple-laddress.cc | 111 poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT); local 114 if (maybe_ne (bytepos, 0)) 115 offset = size_binop (PLUS_EXPR, offset, size_int (bytepos));
|
| tree-affine.cc | 393 poly_int64 bitpos, bitsize, bytepos; local 441 if (!multiple_p (bitpos, BITS_PER_UNIT, &bytepos)) 443 aff_combination_const (comb, type, bytepos);
|
| gimple-ssa-warn-restrict.cc | 496 poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT); local 502 if (bytepos.is_constant (&cstoff))
|
| simplify-rtx.cc | 297 poly_int64 bitsize, bitpos, bytepos, toffset_val = 0; 305 || !multiple_p (bitpos, BITS_PER_UNIT, &bytepos) 309 offset += bytepos + toffset_val; 293 poly_int64 bitsize, bitpos, bytepos, toffset_val = 0; local
|
| stor-layout.cc | 957 tree bytepos; 960 bytepos = TREE_OPERAND (bitpos, 0); 962 bytepos = size_binop (TRUNC_DIV_EXPR, bitpos, bitsize_unit_node); 963 return size_binop (PLUS_EXPR, offset, fold_convert (sizetype, bytepos)); 953 tree bytepos; local
|
| emit-rtl.cc | 2175 poly_int64 bytepos = bits_to_bytes_round_down (apply_bitpos); 2176 attrs.offset -= bytepos; 2178 attrs.size += bytepos; 2167 poly_int64 bytepos = bits_to_bytes_round_down (apply_bitpos); local
|
| gimple-ssa-store-merging.cc | 211 to easily get BASE_ADDR + offset + lowest bytepos; 230 poly_int64 bytepos; member in struct:__anon13746::symbolic_number 371 poly_int64 bitsize, bitpos, bytepos; local 420 if (!multiple_p (bitpos, BITS_PER_UNIT, &bytepos)) 431 n->bytepos = bytepos; 478 if (!(n2->bytepos - n1->bytepos).is_constant (&start2)) 556 n->bytepos = n_start->bytepos; 3733 unsigned HOST_WIDE_INT bytepos = pos \/ BITS_PER_UNIT; local [all...] |
| cfgexpand.cc | 4852 poly_int64 bytepos = bits_to_bytes_round_down (bitpos); local 4853 if (maybe_ne (bytepos, 0)) 4855 op0 = adjust_address_nv (op0, mode1, bytepos); 4880 poly_int64 bytepos; 4881 if (multiple_p (bitpos, BITS_PER_UNIT, &bytepos) 4899 return simplify_gen_subreg (mode, op0, opmode, bytepos);
|
| varasm.cc | 3061 poly_int64 bytepos; 3064 &bytepos)) 3066 offset += bytepos; 3049 poly_int64 bytepos; local
|
| expr.cc | 2995 poly_int64 bytepos = rtx_to_poly_int64 (XEXP (XVECEXP (dst, 0, i), 1)); 3003 gcc_checking_assert (ordered_p (bytepos + bytelen, ssize)); 3004 if (known_size_p (ssize) && maybe_gt (bytepos + bytelen, ssize)) 3016 shift = (bytelen - (ssize - bytepos)) * BITS_PER_UNIT; 3017 bytelen = ssize - bytepos; 3037 && multiple_p (bytepos * BITS_PER_UNIT, GET_MODE_ALIGNMENT (mode)) 3041 emit_move_insn (tmps[i], adjust_address (src, mode, bytepos)); 3055 if (can_div_trunc_p (bytepos, slen0, &elt, &subpos) 3076 gcc_assert (known_eq (bytepos, 0)); 3086 tmps[i] = simplify_gen_subreg (mode, src, GET_MODE (dst), bytepos); 2989 poly_int64 bytepos = rtx_to_poly_int64 (XEXP (XVECEXP (dst, 0, i), 1)); local 3287 poly_int64 bytepos; local 3342 poly_int64 bytepos = rtx_to_poly_int64 (XEXP (XVECEXP (src, 0, i), 1)); local 6132 poly_int64 bytepos; local 7349 poly_int64 bytepos; local 8253 poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT); local 9223 poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT); local 11906 poly_int64 bitsize, bitpos, bytepos; local 12379 poly_int64 bitsize, bitpos, bytepos; local [all...] |
| dwarf2out.cc | 18080 poly_int64 bitsize, bitpos, bytepos; 18089 if (!multiple_p (bitpos, BITS_PER_UNIT, &bytepos)) 18123 if (bytepos.is_constant (&value) && value > 0) 18126 else if (maybe_ne (bytepos, 0)) 18127 loc_list_plus_const (list_ret, bytepos); 19159 poly_int64 bitsize, bitpos, bytepos; 19177 if (!multiple_p (bitpos, BITS_PER_UNIT, &bytepos) 19198 if (bytepos.is_constant (&value) && value > 0) 19201 else if (maybe_ne (bytepos, 0)) 19202 loc_list_plus_const (list_ret, bytepos); 18067 poly_int64 bitsize, bitpos, bytepos; local 19146 poly_int64 bitsize, bitpos, bytepos; local [all...] |
| /src/external/gpl3/gcc.old/dist/gcc/ |
| gimple-laddress.cc | 111 poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT); local 114 if (maybe_ne (bytepos, 0)) 115 offset = size_binop (PLUS_EXPR, offset, size_int (bytepos));
|
| gimple-ssa-warn-restrict.cc | 493 poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT); local 499 if (bytepos.is_constant (&cstoff))
|
| tree-affine.cc | 271 poly_int64 bitpos, bitsize, bytepos; local 393 poly_int64 bitpos, bitsize, bytepos; local 441 if (!multiple_p (bitpos, BITS_PER_UNIT, &bytepos)) 443 aff_combination_const (comb, type, bytepos);
|
| simplify-rtx.cc | 297 poly_int64 bitsize, bitpos, bytepos, toffset_val = 0; 305 || !multiple_p (bitpos, BITS_PER_UNIT, &bytepos) 309 offset += bytepos + toffset_val; 293 poly_int64 bitsize, bitpos, bytepos, toffset_val = 0; local
|
| stor-layout.cc | 929 tree bytepos; 932 bytepos = TREE_OPERAND (bitpos, 0); 934 bytepos = size_binop (TRUNC_DIV_EXPR, bitpos, bitsize_unit_node); 935 return size_binop (PLUS_EXPR, offset, fold_convert (sizetype, bytepos)); 925 tree bytepos; local
|
| cfgexpand.cc | 4820 poly_int64 bytepos = bits_to_bytes_round_down (bitpos); local 4821 if (maybe_ne (bytepos, 0)) 4823 op0 = adjust_address_nv (op0, mode1, bytepos); 4848 poly_int64 bytepos; 4849 if (multiple_p (bitpos, BITS_PER_UNIT, &bytepos) 4867 return simplify_gen_subreg (mode, op0, opmode, bytepos);
|
| emit-rtl.cc | 2159 poly_int64 bytepos = bits_to_bytes_round_down (apply_bitpos); 2160 attrs.offset -= bytepos; 2162 attrs.size += bytepos; 2151 poly_int64 bytepos = bits_to_bytes_round_down (apply_bitpos); local
|
| gimple-ssa-store-merging.cc | 211 to easily get BASE_ADDR + offset + lowest bytepos; 230 poly_int64_pod bytepos; member in struct:__anon16174::symbolic_number 371 poly_int64 bitsize, bitpos, bytepos; local 420 if (!multiple_p (bitpos, BITS_PER_UNIT, &bytepos)) 431 n->bytepos = bytepos; 478 if (!(n2->bytepos - n1->bytepos).is_constant (&start2)) 556 n->bytepos = n_start->bytepos; 3629 unsigned HOST_WIDE_INT bytepos = pos \/ BITS_PER_UNIT; local [all...] |
| varasm.cc | 3027 poly_int64 bytepos; 3030 &bytepos)) 3032 offset += bytepos; 3015 poly_int64 bytepos; local
|
| expr.cc | 2420 poly_int64 bytepos = rtx_to_poly_int64 (XEXP (XVECEXP (dst, 0, i), 1)); 2428 gcc_checking_assert (ordered_p (bytepos + bytelen, ssize)); 2429 if (known_size_p (ssize) && maybe_gt (bytepos + bytelen, ssize)) 2441 shift = (bytelen - (ssize - bytepos)) * BITS_PER_UNIT; 2442 bytelen = ssize - bytepos; 2467 && multiple_p (bytepos * BITS_PER_UNIT, GET_MODE_ALIGNMENT (mode)) 2471 emit_move_insn (tmps[i], adjust_address (src, mode, bytepos)); 2485 if (can_div_trunc_p (bytepos, slen0, &elt, &subpos) 2506 gcc_assert (known_eq (bytepos, 0)); 2516 tmps[i] = simplify_gen_subreg (mode, src, GET_MODE (dst), bytepos); 2414 poly_int64 bytepos = rtx_to_poly_int64 (XEXP (XVECEXP (dst, 0, i), 1)); local 2717 poly_int64 bytepos; local 2775 poly_int64 bytepos = rtx_to_poly_int64 (XEXP (XVECEXP (src, 0, i), 1)); local 5495 poly_int64 bytepos; local 6665 poly_int64 bytepos; local 7539 poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT); local 8510 poly_int64 bytepos = exact_div (bitpos, BITS_PER_UNIT); local 11094 poly_int64 bitsize, bitpos, bytepos; local 11556 poly_int64 bitsize, bitpos, bytepos; local [all...] |
| /src/external/gpl3/gdb/dist/gdb/ |
| amd64-tdep.c | 538 int bytepos = bitpos / 8; 539 if (bytepos % align != 0) 537 int bytepos = bitpos \/ 8; local
|
| /src/external/gpl3/gdb.old/dist/gdb/ |
| amd64-tdep.c | 543 int bytepos = bitpos / 8; 544 if (bytepos % align != 0) 542 int bytepos = bitpos \/ 8; local
|
| /src/external/gpl3/gcc.old/dist/gcc/config/aarch64/ |
| aarch64.cc | 3147 poly_uint64 bytepos; local 3148 if (!wide_bytepos.to_uhwi (&bytepos) 3158 p.offset += bytepos;
|
| /src/external/gpl3/gcc/dist/gcc/config/aarch64/ |
| aarch64.cc | 1140 poly_uint64 bytepos; local 1141 if (!wide_bytepos.to_uhwi (&bytepos) 1151 p.offset += bytepos;
|