HomeSort by: relevance | last modified time | path
    Searched refs:rounded_size (Results 1 - 25 of 39) sorted by relevancy

1 2

  /src/external/gpl3/gcc.old/dist/libobjc/
sarray.c 296 size_t rounded_size = (new_max_index + 1) * INDEX_CAPACITY;
304 size_t rounded_size = (new_max_index + 1) * BUCKET_SIZE;
316 if (rounded_size <= array->capacity)
323 if (rounded_size > array->capacity)
327 rounded_size = (new_max_index + 1) * INDEX_CAPACITY;
330 rounded_size = (new_max_index + 1) * BUCKET_SIZE;
334 array->capacity = rounded_size;
295 size_t rounded_size = (new_max_index + 1) * INDEX_CAPACITY; local
  /src/external/gpl3/gcc/dist/libobjc/
sarray.c 296 size_t rounded_size = (new_max_index + 1) * INDEX_CAPACITY;
304 size_t rounded_size = (new_max_index + 1) * BUCKET_SIZE;
316 if (rounded_size <= array->capacity)
323 if (rounded_size > array->capacity)
327 rounded_size = (new_max_index + 1) * INDEX_CAPACITY;
330 rounded_size = (new_max_index + 1) * BUCKET_SIZE;
334 array->capacity = rounded_size;
295 size_t rounded_size = (new_max_index + 1) * INDEX_CAPACITY; local
  /src/external/gpl3/gcc.old/dist/gcc/
explow.cc 1781 rtx rounded_size, rounded_size_op, test_addr, last_addr, temp;
1787 /* ROUNDED_SIZE = SIZE & -PROBE_INTERVAL */
1788 rounded_size
1791 rounded_size_op = force_operand (rounded_size, NULL_RTX);
1802 /* LAST_ADDR = SP + FIRST + ROUNDED_SIZE. */
1817 until it is equal to ROUNDED_SIZE. */
1841 that SIZE is equal to ROUNDED_SIZE. */
1843 /* TEMP = SIZE - ROUNDED_SIZE. */
1844 temp = simplify_gen_binary (MINUS, Pmode, size, rounded_size);
1877 We compute ROUNDED_SIZE, LAST_ADDR, RESIDUAL and PROBE_INTERVAL
1772 rtx rounded_size, rounded_size_op, test_addr, last_addr, temp; local
1998 rtx rounded_size, last_addr, residual; local
2168 rtx rounded_size, rounded_size_op, last_addr, temp; local
    [all...]
targhooks.cc 2346 tree addr, t, type_size, rounded_size, valist_tmp; local
2423 rounded_size = round_up (type_size, align);
2425 /* Reduce rounded_size so it's sharable with the postqueue. */
2426 gimplify_expr (&rounded_size, pre_p, post_p, is_gimple_val, fb_rvalue);
2430 if (PAD_VARARGS_DOWN && !integer_zerop (rounded_size))
2434 rounded_size, size_int (align));
2436 size_binop (MINUS_EXPR, rounded_size, type_size));
2441 t = fold_build_pointer_plus (valist_tmp, rounded_size);
function.cc 851 poly_int64 rounded_size = aligned_upper_bound (size, alignment);
853 if (known_ge (best_p->size - rounded_size, alignment))
857 p->size = best_p->size - rounded_size;
858 p->base_offset = best_p->base_offset + rounded_size;
859 p->full_size = best_p->full_size - rounded_size;
860 p->slot = adjust_address_nv (best_p->slot, BLKmode, rounded_size);
867 best_p->size = rounded_size;
868 best_p->full_size = rounded_size;
846 poly_int64 rounded_size = aligned_upper_bound (size, alignment); local
recog.cc 1758 poly_int64 rounded_size = GET_MODE_SIZE (mode);
1761 rounded_size = PUSH_ROUNDING (MACRO_INT (rounded_size));
1766 if (known_eq (rounded_size, GET_MODE_SIZE (mode)))
1779 ? maybe_ne (offset, -rounded_size)
1780 : maybe_ne (offset, rounded_size)))
  /src/external/gpl3/gcc/dist/gcc/
explow.cc 1821 rtx rounded_size, rounded_size_op, test_addr, last_addr, temp;
1827 /* ROUNDED_SIZE = SIZE & -PROBE_INTERVAL */
1828 rounded_size
1831 rounded_size_op = force_operand (rounded_size, NULL_RTX);
1842 /* LAST_ADDR = SP + FIRST + ROUNDED_SIZE. */
1857 until it is equal to ROUNDED_SIZE. */
1884 that SIZE is equal to ROUNDED_SIZE. */
1886 /* TEMP = SIZE - ROUNDED_SIZE. */
1887 temp = simplify_gen_binary (MINUS, Pmode, size, rounded_size);
1920 We compute ROUNDED_SIZE, LAST_ADDR, RESIDUAL and PROBE_INTERVAL
1812 rtx rounded_size, rounded_size_op, test_addr, last_addr, temp; local
2041 rtx rounded_size, last_addr, residual; local
2211 rtx rounded_size, rounded_size_op, last_addr, temp; local
    [all...]
targhooks.cc 2492 tree addr, t, type_size, rounded_size, valist_tmp; local
2569 rounded_size = round_up (type_size, align);
2571 /* Reduce rounded_size so it's sharable with the postqueue. */
2572 gimplify_expr (&rounded_size, pre_p, post_p, is_gimple_val, fb_rvalue);
2576 if (PAD_VARARGS_DOWN && !integer_zerop (rounded_size))
2580 rounded_size, size_int (align));
2582 size_binop (MINUS_EXPR, rounded_size, type_size));
2587 t = fold_build_pointer_plus (valist_tmp, rounded_size);
function.cc 853 poly_int64 rounded_size = aligned_upper_bound (size, alignment);
855 if (known_ge (best_p->size - rounded_size, alignment))
859 p->size = best_p->size - rounded_size;
860 p->base_offset = best_p->base_offset + rounded_size;
861 p->full_size = best_p->full_size - rounded_size;
862 p->slot = adjust_address_nv (best_p->slot, BLKmode, rounded_size);
869 best_p->size = rounded_size;
870 best_p->full_size = rounded_size;
848 poly_int64 rounded_size = aligned_upper_bound (size, alignment); local
recog.cc 1771 poly_int64 rounded_size = GET_MODE_SIZE (mode);
1774 rounded_size = PUSH_ROUNDING (MACRO_INT (rounded_size));
1779 if (known_eq (rounded_size, GET_MODE_SIZE (mode)))
1792 ? maybe_ne (offset, -rounded_size)
1793 : maybe_ne (offset, rounded_size)))
  /src/external/gpl3/gcc.old/dist/gcc/config/rs6000/
rs6000-logue.cc 1573 HOST_WIDE_INT rounded_size = ROUND_DOWN (orig_size, probe_interval);
1579 if (rounded_size != orig_size
1580 || rounded_size > probe_interval
1602 if (rounded_size == probe_interval)
1606 dump_stack_clash_frame_info (PROBE_INLINE, rounded_size != orig_size);
1608 else if (rounded_size <= 8 * probe_interval)
1616 for (int i = 0; i < rounded_size; i += probe_interval)
1626 dump_stack_clash_frame_info (PROBE_INLINE, rounded_size != orig_size);
1633 rtx rs = GEN_INT (-rounded_size);
1683 dump_stack_clash_frame_info (PROBE_LOOP, rounded_size != orig_size)
1569 HOST_WIDE_INT rounded_size = ROUND_DOWN (orig_size, probe_interval); local
1821 HOST_WIDE_INT rounded_size; local
    [all...]
  /src/external/gpl3/gcc/dist/gcc/config/rs6000/
rs6000-logue.cc 1566 HOST_WIDE_INT rounded_size = ROUND_DOWN (orig_size, probe_interval);
1572 if (rounded_size != orig_size
1573 || rounded_size > probe_interval
1595 if (rounded_size == probe_interval)
1599 dump_stack_clash_frame_info (PROBE_INLINE, rounded_size != orig_size);
1601 else if (rounded_size <= 8 * probe_interval)
1609 for (int i = 0; i < rounded_size; i += probe_interval)
1619 dump_stack_clash_frame_info (PROBE_INLINE, rounded_size != orig_size);
1626 rtx rs = GEN_INT (-rounded_size);
1676 dump_stack_clash_frame_info (PROBE_LOOP, rounded_size != orig_size)
1562 HOST_WIDE_INT rounded_size = ROUND_DOWN (orig_size, probe_interval); local
1814 HOST_WIDE_INT rounded_size; local
    [all...]
  /src/external/gpl3/gcc.old/dist/gcc/config/loongarch/
loongarch.cc 1152 HOST_WIDE_INT rounded_size; local
1163 rounded_size = ROUND_DOWN (size, PROBE_INTERVAL);
1179 /* LAST_ADDR = SP + FIRST + ROUNDED_SIZE. */
1180 if (rounded_size == 0)
1184 emit_move_insn (r12, GEN_INT (rounded_size));
1196 until it is equal to ROUNDED_SIZE. */
1202 that SIZE is equal to ROUNDED_SIZE. */
1204 if (size != rounded_size)
1207 emit_stack_probe (plus_constant (Pmode, r12, rounded_size - size));
1211 for (i = 2048; i < (size - rounded_size); i += 2048
    [all...]
  /src/external/gpl3/gcc.old/dist/gcc/config/msp430/
msp430.cc 829 tree addr, t, type_size, rounded_size, valist_tmp;
888 rounded_size = round_up (type_size, align);
890 /* Reduce rounded_size so it's sharable with the postqueue. */
891 gimplify_expr (&rounded_size, pre_p, post_p, is_gimple_val, fb_rvalue);
897 t = fold_build_pointer_plus (valist_tmp, rounded_size);
823 tree addr, t, type_size, rounded_size, valist_tmp; local
  /src/external/gpl3/gcc/dist/gcc/config/msp430/
msp430.cc 829 tree addr, t, type_size, rounded_size, valist_tmp;
888 rounded_size = round_up (type_size, align);
890 /* Reduce rounded_size so it's sharable with the postqueue. */
891 gimplify_expr (&rounded_size, pre_p, post_p, is_gimple_val, fb_rvalue);
897 t = fold_build_pointer_plus (valist_tmp, rounded_size);
823 tree addr, t, type_size, rounded_size, valist_tmp; local
  /src/sys/external/bsd/compiler_rt/dist/lib/asan/
asan_allocator.cc 427 uptr rounded_size = RoundUpTo(Max(size, kChunkHeader2Size), alignment); local
428 uptr needed_size = rounded_size + rz_size;
  /src/external/gpl3/gcc.old/dist/libsanitizer/asan/
asan_allocator.cpp 511 uptr rounded_size = RoundUpTo(Max(size, kChunkHeader2Size), alignment); local
512 uptr needed_size = rounded_size + rz_size;
  /src/external/gpl3/gcc/dist/libsanitizer/asan/
asan_allocator.cpp 556 uptr rounded_size = RoundUpTo(Max(size, kChunkHeader2Size), alignment); local
557 uptr needed_size = rounded_size + rz_size;
  /src/external/gpl3/gcc.old/dist/gcc/config/rl78/
rl78.cc 1761 int rounded_size;
1764 rounded_size = arg.promoted_size_in_bytes ();
1765 if (rounded_size & 1)
1766 rounded_size ++;
1767 (*cum) += rounded_size;
1756 int rounded_size; local
  /src/external/gpl3/gcc/dist/gcc/config/rl78/
rl78.cc 1761 int rounded_size;
1764 rounded_size = arg.promoted_size_in_bytes ();
1765 if (rounded_size & 1)
1766 rounded_size ++;
1767 (*cum) += rounded_size;
1756 int rounded_size; local
  /src/external/gpl3/gcc.old/dist/gcc/config/rx/
rx.cc 1081 unsigned int rounded_size;
1088 rounded_size = rx_round_up (size, UNITS_PER_WORD);
1092 if (rounded_size + bytes_so_far > MAX_NUM_ARG_BYTES)
1077 unsigned int rounded_size; local
  /src/external/gpl3/gcc/dist/gcc/config/loongarch/
loongarch.cc 1149 HOST_WIDE_INT rounded_size; local
1167 rounded_size = ROUND_DOWN (size, interval);
1173 /* If rounded_size is zero, it means that the space requested by
1176 if (rounded_size != 0)
1188 until it is equal to ROUNDED_SIZE. */
1190 if (rounded_size <= STACK_CLASH_MAX_UNROLL_PAGES * interval)
1192 for (HOST_WIDE_INT i = 0; i < rounded_size; i += interval)
1205 dump_stack_clash_frame_info (PROBE_INLINE, size != rounded_size);
1209 emit_move_insn (r12, GEN_INT (rounded_size));
1218 dump_stack_clash_frame_info (PROBE_LOOP, size != rounded_size);
    [all...]
  /src/external/gpl3/gcc/dist/gcc/config/rx/
rx.cc 1082 unsigned int rounded_size;
1089 rounded_size = rx_round_up (size, UNITS_PER_WORD);
1093 if (rounded_size + bytes_so_far > MAX_NUM_ARG_BYTES)
1078 unsigned int rounded_size; local
  /src/external/gpl3/gcc.old/dist/gcc/config/i386/
i386.cc 7823 HOST_WIDE_INT rounded_size = size & -probe_interval;
7827 rtx addr = plus_constant (Pmode, stack_pointer_rtx, -rounded_size);
7833 emit_move_insn (sr.reg, GEN_INT (-rounded_size));
7842 m->fs.cfa_offset + rounded_size));
7847 rtx size_rtx = GEN_INT (rounded_size);
7852 m->fs.cfa_offset += rounded_size;
7858 m->fs.sp_offset += rounded_size;
7862 is equal to ROUNDED_SIZE. */
7864 if (size != rounded_size)
7867 GEN_INT (rounded_size - size), -1
7806 HOST_WIDE_INT rounded_size = size & -probe_interval; local
7954 HOST_WIDE_INT rounded_size, last; local
    [all...]
  /src/external/gpl3/gcc.old/dist/gcc/config/mips/
mips.cc 6945 tree addr, t, type_size, rounded_size, valist_tmp;
6997 rounded_size = round_up (type_size, align);
6999 /* Reduce rounded_size so it's sharable with the postqueue. */
7000 gimplify_expr (&rounded_size, pre_p, post_p, is_gimple_val, fb_rvalue);
7004 if (PAD_VARARGS_DOWN && !integer_zerop (rounded_size))
7008 rounded_size, size_int (align));
7010 size_binop (MINUS_EXPR, rounded_size, type_size));
7015 t = fold_build_pointer_plus (valist_tmp, rounded_size);
12202 HOST_WIDE_INT rounded_size;
12212 rounded_size = ROUND_DOWN (size, PROBE_INTERVAL)
6914 tree addr, t, type_size, rounded_size, valist_tmp; local
12151 HOST_WIDE_INT rounded_size; local
    [all...]

Completed in 110 milliseconds

1 2