| /src/external/bsd/jemalloc/dist/test/stress/ |
| batch_alloc.c | 29 batch_alloc_wrapper(size_t batch) { 31 {batch_ptrs + batch_ptrs_next, batch, SIZE, 0}; 36 assert_zu_eq(filled, batch, ""); 40 item_alloc_wrapper(size_t batch) { 41 for (size_t i = item_ptrs_next, end = i + batch; i < end; ++i) { 57 batch_alloc_without_free(size_t batch) { 58 batch_alloc_wrapper(batch); 59 batch_ptrs_next += batch; 63 item_alloc_without_free(size_t batch) { 64 item_alloc_wrapper(batch); [all...] |
| /src/external/lgpl2/userspace-rcu/dist/tests/benchmark/ |
| runpaul-phase1.sh | 35 echo Executing batch RCU test 42 rm -f batch-rcu.log 46 echo "./runtests-batch.sh ${NR_READERS} ${NR_WRITERS} ${DURATION} -d 0 -b ${BATCH_SIZE} ${EXTRA_OPTS} | tee -a batch-rcu.log" >> runall.log 47 ./runtests-batch.sh ${NR_READERS} ${NR_WRITERS} ${DURATION} -d 0 -b ${BATCH_SIZE} ${EXTRA_OPTS} | tee -a batch-rcu.log
|
| subphase4.sh | 35 echo Executing batch RCU test 42 rm -f batch-rcu.log 46 echo "./runtests-batch.sh ${NR_READERS} ${NR_WRITERS} ${DURATION} -d 0 -b ${BATCH_SIZE} ${EXTRA_OPTS} | tee -a batch-rcu.log" >> runall.log 47 ./runtests-batch.sh ${NR_READERS} ${NR_WRITERS} ${DURATION} -d 0 -b ${BATCH_SIZE} ${EXTRA_OPTS} | tee -a batch-rcu.log
|
| /src/external/bsd/jemalloc/dist/test/unit/ |
| batch_alloc.c | 9 verify_batch_basic(tsd_t *tsd, void **ptrs, size_t batch, size_t usize, 11 for (size_t i = 0; i < batch; ++i) { 23 verify_batch_locality(tsd_t *tsd, void **ptrs, size_t batch, size_t usize, 27 * Checking batch locality when prof is on is feasible but 33 for (size_t i = 0, j = 0; i < batch; ++i, ++j) { 37 if (j == 0 && batch - i < nregs) { 54 release_batch(void **ptrs, size_t batch, size_t size) { 55 for (size_t i = 0; i < batch; ++i) { 123 size_t batch = base + (size_t)j; local 124 assert(batch < BATCH_MAX) [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
| igt_spinner.c | 52 spin->batch = vaddr; 102 u32 *batch; local 140 batch = spin->batch; 143 *batch++ = MI_STORE_DWORD_IMM_GEN4; 144 *batch++ = lower_32_bits(hws_address(hws, rq)); 145 *batch++ = upper_32_bits(hws_address(hws, rq)); 147 *batch++ = MI_STORE_DWORD_IMM_GEN4; 148 *batch++ = 0; 149 *batch++ = hws_address(hws, rq) [all...] |
| igt_spinner.h | 25 u32 *batch; member in struct:igt_spinner
|
| i915_request.c | 537 * (individually), and wait for the batch to complete. We can check 654 struct i915_vma *batch) 664 batch->node.start, 665 batch->node.size, 681 struct i915_vma *batch; local 686 * (individually), and wait for the batch to complete. We can check 690 batch = empty_batch(i915); 691 if (IS_ERR(batch)) 692 return PTR_ERR(batch); 707 request = empty_request(engine, batch); 831 struct i915_vma *batch; local 965 struct i915_vma *batch; local [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
| igt_gem_utils.c | 117 struct i915_vma *batch; local 124 batch = igt_emit_store_dw(vma, offset, count, val); 125 if (IS_ERR(batch)) 126 return PTR_ERR(batch); 139 batch->node.start, batch->node.size, 144 i915_vma_lock(batch); 145 err = i915_request_await_object(rq, batch->obj, false); 147 err = i915_vma_move_to_active(batch, rq, 0); 148 i915_vma_unlock(batch); [all...] |
| /src/usr.bin/at/ |
| Makefile | 9 ${BINDIR}/at ${BINDIR}/batch 10 MLINKS= at.1 batch.1 \
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
| i915_gem_object_blt.c | 27 struct i915_vma *batch; local 90 batch = i915_vma_instance(pool->obj, ce->vm, NULL); 91 if (IS_ERR(batch)) { 92 err = PTR_ERR(batch); 96 err = i915_vma_pin(batch, 0, 0, PIN_USER); 100 batch->private = pool; 101 return batch; 137 struct i915_vma *batch; local 155 batch = intel_emit_vma_fill_blt(ce, vma, value); 156 if (IS_ERR(batch)) { 211 struct i915_vma *batch; local 325 struct i915_vma *vma[2], *batch; local [all...] |
| i915_gem_client_blt.c | 166 struct i915_vma *batch; local 184 batch = intel_emit_vma_fill_blt(w->ce, vma, w->value); 185 if (IS_ERR(batch)) { 186 err = PTR_ERR(batch); 201 err = intel_emit_vma_mark_active(batch, rq); 221 batch->node.start, batch->node.size, 231 intel_emit_vma_release(w->ce, batch);
|
| i915_gem_execbuffer.c | 147 * Any render targets written to in the batch must be flagged with 218 * Before any batch is given extra privileges we first must check that it 240 struct i915_vma *batch; /** identity of the batch obj/vma */ member in struct:i915_execbuffer 275 u32 batch_start_offset; /** Location within object of batch */ 276 u32 batch_len; /** Length of batch within object */ 519 * SNA is doing fancy tricks with compressing batch buffers, which leads 521 * relocate address is still positive, except when the batch is placed 534 eb->batch = vma; 552 eb->batch = NULL 743 unsigned int i, batch; local 1155 struct i915_vma *batch; local 1276 u32 *batch; local 2006 struct i915_vma *batch; member in struct:eb_parse_work [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
| intel_renderstate.h | 37 const u32 *batch; member in struct:intel_renderstate_rodata 44 .batch = gen ## _g ## _null_state_batch, \
|
| selftest_hangcheck.c | 57 u32 *batch; member in struct:hang 100 h->batch = vaddr; 147 u32 *batch; local 167 h->batch = vaddr; 205 batch = h->batch; 207 *batch++ = MI_STORE_DWORD_IMM_GEN4; 208 *batch++ = lower_32_bits(hws_address(hws, rq)); 209 *batch++ = upper_32_bits(hws_address(hws, rq)); 210 *batch++ = rq->fence.seqno [all...] |
| /src/usr.bin/mail/ |
| mime_detach.c | 62 int batch; member in struct:__anon8280 71 detach_ctl.batch = value(ENAME_MIME_DETACH_BATCH) != NULL; 72 detach_ctl.ask = detach_ctl.batch ? 0 : 1; 97 if (!detach_ctl.batch) { 146 detach_ctl.batch = 1; 155 detach_ctl.batch = 1; 237 detach_ctl.batch = 0; 242 } while (!detach_ctl.batch);
|
| /src/sys/external/bsd/drm2/include/linux/ |
| shrinker.h | 49 size_t batch; member in struct:shrinker
|
| /src/external/mpl/bind/dist/tests/ |
| unit-test-driver.sh.in | 45 BINARY=$(gdb --batch --core="${CORE_DUMP}" 2>/dev/null | sed -n "s/^Core was generated by \`\(.*\)'\.\$/\1/p") 52 --batch \
|
| /src/tests/usr.bin/gdb/ |
| t_regress.sh | 48 gdb --batch -x test.gdb dig >gdb.out 68 gdb --batch -x test.gdb ./test >gdb.out 2>&1 87 gdb --batch -x test.gdb >gdb.out 2>&1
|
| /src/external/mpl/bind/dist/bin/tests/system/ |
| get_core_dumps.sh | 33 binary=$(gdb --batch --core="$coredump" 2>/dev/null | sed -ne "s|Core was generated by \`\([^' ]*\)[' ].*|\1|p") 38 -batch \ 47 -batch \
|
| /src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
| vmwgfx_mob.c | 241 struct vmw_otable_batch *batch) 245 struct vmw_otable *otables = batch->otables; 254 for (i = 0; i < batch->num_otables; ++i) { 266 0, false, &batch->otable_bo); 271 ret = ttm_bo_reserve(batch->otable_bo, false, true, NULL); 273 ret = vmw_bo_driver.ttm_tt_populate(batch->otable_bo->ttm, &ctx); 276 ret = vmw_bo_map_dma(batch->otable_bo); 280 ttm_bo_unreserve(batch->otable_bo); 283 for (i = 0; i < batch->num_otables; ++i) { 284 if (!batch->otables[i].enabled [all...] |
| /src/external/bsd/openldap/dist/tests/data/tls/ |
| create-crt.sh | 80 -batch > /dev/null 2>&1 84 -batch >/dev/null 2>&1 96 -batch >/dev/null 2>&1 100 -cert ca/certs/testsuiteCA.crt -batch >/dev/null 2>&1 112 -batch > /dev/null 2>&1 116 -batch >/dev/null 2>&1
|
| /src/external/gpl3/autoconf/dist/build-aux/ |
| elisp-comp | 79 $EMACS -batch -q -l script -f batch-byte-compile *.el || exit $?
|
| /src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
| tsan_dense_alloc.h | 112 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); local 116 new(batch + i) T; 117 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size; 119 *(IndexT*)(batch + kL2Size - 1) = 0; 121 map_[fillpos_++] = batch;
|
| /src/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
| tsan_dense_alloc.h | 125 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); local 129 new(batch + i) T; 130 *(IndexT *)(batch + i) = i + 1 + fillpos * kL2Size; 132 *(IndexT*)(batch + kL2Size - 1) = 0; 134 map_[fillpos] = batch;
|
| /src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/ |
| sanitizer_quarantine.h | 33 void *batch[kSize]; member in struct:__sanitizer::QuarantineBatch 37 batch[0] = ptr; 38 this->size = size + sizeof(QuarantineBatch); // Account for the batch size. 41 // The total size of quarantined nodes recorded in this batch. 48 batch[count++] = ptr; 61 batch[count + i] = from->batch[i]; 166 // require some tuning). It saves us merge attempt when the batch list 187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); 189 PREFETCH(b->batch[i]) [all...] |