HomeSort by: relevance | last modified time | path
    Searched refs:batch (Results 1 - 25 of 119) sorted by relevancy

1 2 3 4 5

  /src/external/bsd/jemalloc/dist/test/stress/
batch_alloc.c 29 batch_alloc_wrapper(size_t batch) {
31 {batch_ptrs + batch_ptrs_next, batch, SIZE, 0};
36 assert_zu_eq(filled, batch, "");
40 item_alloc_wrapper(size_t batch) {
41 for (size_t i = item_ptrs_next, end = i + batch; i < end; ++i) {
57 batch_alloc_without_free(size_t batch) {
58 batch_alloc_wrapper(batch);
59 batch_ptrs_next += batch;
63 item_alloc_without_free(size_t batch) {
64 item_alloc_wrapper(batch);
    [all...]
  /src/external/lgpl2/userspace-rcu/dist/tests/benchmark/
runpaul-phase1.sh 35 echo Executing batch RCU test
42 rm -f batch-rcu.log
46 echo "./runtests-batch.sh ${NR_READERS} ${NR_WRITERS} ${DURATION} -d 0 -b ${BATCH_SIZE} ${EXTRA_OPTS} | tee -a batch-rcu.log" >> runall.log
47 ./runtests-batch.sh ${NR_READERS} ${NR_WRITERS} ${DURATION} -d 0 -b ${BATCH_SIZE} ${EXTRA_OPTS} | tee -a batch-rcu.log
subphase4.sh 35 echo Executing batch RCU test
42 rm -f batch-rcu.log
46 echo "./runtests-batch.sh ${NR_READERS} ${NR_WRITERS} ${DURATION} -d 0 -b ${BATCH_SIZE} ${EXTRA_OPTS} | tee -a batch-rcu.log" >> runall.log
47 ./runtests-batch.sh ${NR_READERS} ${NR_WRITERS} ${DURATION} -d 0 -b ${BATCH_SIZE} ${EXTRA_OPTS} | tee -a batch-rcu.log
  /src/external/bsd/jemalloc/dist/test/unit/
batch_alloc.c 9 verify_batch_basic(tsd_t *tsd, void **ptrs, size_t batch, size_t usize,
11 for (size_t i = 0; i < batch; ++i) {
23 verify_batch_locality(tsd_t *tsd, void **ptrs, size_t batch, size_t usize,
27 * Checking batch locality when prof is on is feasible but
33 for (size_t i = 0, j = 0; i < batch; ++i, ++j) {
37 if (j == 0 && batch - i < nregs) {
54 release_batch(void **ptrs, size_t batch, size_t size) {
55 for (size_t i = 0; i < batch; ++i) {
123 size_t batch = base + (size_t)j; local
124 assert(batch < BATCH_MAX)
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/selftests/
igt_spinner.c 52 spin->batch = vaddr;
102 u32 *batch; local
140 batch = spin->batch;
143 *batch++ = MI_STORE_DWORD_IMM_GEN4;
144 *batch++ = lower_32_bits(hws_address(hws, rq));
145 *batch++ = upper_32_bits(hws_address(hws, rq));
147 *batch++ = MI_STORE_DWORD_IMM_GEN4;
148 *batch++ = 0;
149 *batch++ = hws_address(hws, rq)
    [all...]
igt_spinner.h 25 u32 *batch; member in struct:igt_spinner
i915_request.c 537 * (individually), and wait for the batch to complete. We can check
654 struct i915_vma *batch)
664 batch->node.start,
665 batch->node.size,
681 struct i915_vma *batch; local
686 * (individually), and wait for the batch to complete. We can check
690 batch = empty_batch(i915);
691 if (IS_ERR(batch))
692 return PTR_ERR(batch);
707 request = empty_request(engine, batch);
831 struct i915_vma *batch; local
965 struct i915_vma *batch; local
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/
igt_gem_utils.c 117 struct i915_vma *batch; local
124 batch = igt_emit_store_dw(vma, offset, count, val);
125 if (IS_ERR(batch))
126 return PTR_ERR(batch);
139 batch->node.start, batch->node.size,
144 i915_vma_lock(batch);
145 err = i915_request_await_object(rq, batch->obj, false);
147 err = i915_vma_move_to_active(batch, rq, 0);
148 i915_vma_unlock(batch);
    [all...]
  /src/usr.bin/at/
Makefile 9 ${BINDIR}/at ${BINDIR}/batch
10 MLINKS= at.1 batch.1 \
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/
i915_gem_object_blt.c 27 struct i915_vma *batch; local
90 batch = i915_vma_instance(pool->obj, ce->vm, NULL);
91 if (IS_ERR(batch)) {
92 err = PTR_ERR(batch);
96 err = i915_vma_pin(batch, 0, 0, PIN_USER);
100 batch->private = pool;
101 return batch;
137 struct i915_vma *batch; local
155 batch = intel_emit_vma_fill_blt(ce, vma, value);
156 if (IS_ERR(batch)) {
211 struct i915_vma *batch; local
325 struct i915_vma *vma[2], *batch; local
    [all...]
i915_gem_client_blt.c 166 struct i915_vma *batch; local
184 batch = intel_emit_vma_fill_blt(w->ce, vma, w->value);
185 if (IS_ERR(batch)) {
186 err = PTR_ERR(batch);
201 err = intel_emit_vma_mark_active(batch, rq);
221 batch->node.start, batch->node.size,
231 intel_emit_vma_release(w->ce, batch);
i915_gem_execbuffer.c 147 * Any render targets written to in the batch must be flagged with
218 * Before any batch is given extra privileges we first must check that it
240 struct i915_vma *batch; /** identity of the batch obj/vma */ member in struct:i915_execbuffer
275 u32 batch_start_offset; /** Location within object of batch */
276 u32 batch_len; /** Length of batch within object */
519 * SNA is doing fancy tricks with compressing batch buffers, which leads
521 * relocate address is still positive, except when the batch is placed
534 eb->batch = vma;
552 eb->batch = NULL
743 unsigned int i, batch; local
1155 struct i915_vma *batch; local
1276 u32 *batch; local
2006 struct i915_vma *batch; member in struct:eb_parse_work
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/gt/
intel_renderstate.h 37 const u32 *batch; member in struct:intel_renderstate_rodata
44 .batch = gen ## _g ## _null_state_batch, \
selftest_hangcheck.c 57 u32 *batch; member in struct:hang
100 h->batch = vaddr;
147 u32 *batch; local
167 h->batch = vaddr;
205 batch = h->batch;
207 *batch++ = MI_STORE_DWORD_IMM_GEN4;
208 *batch++ = lower_32_bits(hws_address(hws, rq));
209 *batch++ = upper_32_bits(hws_address(hws, rq));
210 *batch++ = rq->fence.seqno
    [all...]
  /src/usr.bin/mail/
mime_detach.c 62 int batch; member in struct:__anon8280
71 detach_ctl.batch = value(ENAME_MIME_DETACH_BATCH) != NULL;
72 detach_ctl.ask = detach_ctl.batch ? 0 : 1;
97 if (!detach_ctl.batch) {
146 detach_ctl.batch = 1;
155 detach_ctl.batch = 1;
237 detach_ctl.batch = 0;
242 } while (!detach_ctl.batch);
  /src/sys/external/bsd/drm2/include/linux/
shrinker.h 49 size_t batch; member in struct:shrinker
  /src/external/mpl/bind/dist/tests/
unit-test-driver.sh.in 45 BINARY=$(gdb --batch --core="${CORE_DUMP}" 2>/dev/null | sed -n "s/^Core was generated by \`\(.*\)'\.\$/\1/p")
52 --batch \
  /src/tests/usr.bin/gdb/
t_regress.sh 48 gdb --batch -x test.gdb dig >gdb.out
68 gdb --batch -x test.gdb ./test >gdb.out 2>&1
87 gdb --batch -x test.gdb >gdb.out 2>&1
  /src/external/mpl/bind/dist/bin/tests/system/
get_core_dumps.sh 33 binary=$(gdb --batch --core="$coredump" 2>/dev/null | sed -ne "s|Core was generated by \`\([^' ]*\)[' ].*|\1|p")
38 -batch \
47 -batch \
  /src/sys/external/bsd/drm2/dist/drm/vmwgfx/
vmwgfx_mob.c 241 struct vmw_otable_batch *batch)
245 struct vmw_otable *otables = batch->otables;
254 for (i = 0; i < batch->num_otables; ++i) {
266 0, false, &batch->otable_bo);
271 ret = ttm_bo_reserve(batch->otable_bo, false, true, NULL);
273 ret = vmw_bo_driver.ttm_tt_populate(batch->otable_bo->ttm, &ctx);
276 ret = vmw_bo_map_dma(batch->otable_bo);
280 ttm_bo_unreserve(batch->otable_bo);
283 for (i = 0; i < batch->num_otables; ++i) {
284 if (!batch->otables[i].enabled
    [all...]
  /src/external/bsd/openldap/dist/tests/data/tls/
create-crt.sh 80 -batch > /dev/null 2>&1
84 -batch >/dev/null 2>&1
96 -batch >/dev/null 2>&1
100 -cert ca/certs/testsuiteCA.crt -batch >/dev/null 2>&1
112 -batch > /dev/null 2>&1
116 -batch >/dev/null 2>&1
  /src/external/gpl3/autoconf/dist/build-aux/
elisp-comp 79 $EMACS -batch -q -l script -f batch-byte-compile *.el || exit $?
  /src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/
tsan_dense_alloc.h 112 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); local
116 new(batch + i) T;
117 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size;
119 *(IndexT*)(batch + kL2Size - 1) = 0;
121 map_[fillpos_++] = batch;
  /src/external/gpl3/gcc.old/dist/libsanitizer/tsan/
tsan_dense_alloc.h 125 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); local
129 new(batch + i) T;
130 *(IndexT *)(batch + i) = i + 1 + fillpos * kL2Size;
132 *(IndexT*)(batch + kL2Size - 1) = 0;
134 map_[fillpos] = batch;
  /src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/
sanitizer_quarantine.h 33 void *batch[kSize]; member in struct:__sanitizer::QuarantineBatch
37 batch[0] = ptr;
38 this->size = size + sizeof(QuarantineBatch); // Account for the batch size.
41 // The total size of quarantined nodes recorded in this batch.
48 batch[count++] = ptr;
61 batch[count + i] = from->batch[i];
166 // require some tuning). It saves us merge attempt when the batch list
187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch));
189 PREFETCH(b->batch[i])
    [all...]

Completed in 189 milliseconds

1 2 3 4 5