HomeSort by: relevance | last modified time | path
    Searched defs:batch (Results 1 - 25 of 49) sorted by relevancy

1 2

  /src/sys/external/bsd/drm2/dist/drm/i915/selftests/
igt_spinner.h 25 u32 *batch; member in struct:igt_spinner
igt_spinner.c 52 spin->batch = vaddr;
102 u32 *batch; local
140 batch = spin->batch;
143 *batch++ = MI_STORE_DWORD_IMM_GEN4;
144 *batch++ = lower_32_bits(hws_address(hws, rq));
145 *batch++ = upper_32_bits(hws_address(hws, rq));
147 *batch++ = MI_STORE_DWORD_IMM_GEN4;
148 *batch++ = 0;
149 *batch++ = hws_address(hws, rq)
    [all...]
i915_request.c 537 * (individually), and wait for the batch to complete. We can check
654 struct i915_vma *batch)
664 batch->node.start,
665 batch->node.size,
681 struct i915_vma *batch; local
686 * (individually), and wait for the batch to complete. We can check
690 batch = empty_batch(i915);
691 if (IS_ERR(batch))
692 return PTR_ERR(batch);
707 request = empty_request(engine, batch);
831 struct i915_vma *batch; local
965 struct i915_vma *batch; local
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/
igt_gem_utils.c 117 struct i915_vma *batch; local
124 batch = igt_emit_store_dw(vma, offset, count, val);
125 if (IS_ERR(batch))
126 return PTR_ERR(batch);
139 batch->node.start, batch->node.size,
144 i915_vma_lock(batch);
145 err = i915_request_await_object(rq, batch->obj, false);
147 err = i915_vma_move_to_active(batch, rq, 0);
148 i915_vma_unlock(batch);
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/gt/
intel_renderstate.h 37 const u32 *batch; member in struct:intel_renderstate_rodata
44 .batch = gen ## _g ## _null_state_batch, \
  /src/sys/external/bsd/drm2/include/linux/
shrinker.h 49 size_t batch; member in struct:shrinker
  /src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/
tsan_dense_alloc.h 112 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); local
116 new(batch + i) T;
117 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size;
119 *(IndexT*)(batch + kL2Size - 1) = 0;
121 map_[fillpos_++] = batch;
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/
i915_gem_object_blt.c 27 struct i915_vma *batch; local
90 batch = i915_vma_instance(pool->obj, ce->vm, NULL);
91 if (IS_ERR(batch)) {
92 err = PTR_ERR(batch);
96 err = i915_vma_pin(batch, 0, 0, PIN_USER);
100 batch->private = pool;
101 return batch;
137 struct i915_vma *batch; local
155 batch = intel_emit_vma_fill_blt(ce, vma, value);
156 if (IS_ERR(batch)) {
211 struct i915_vma *batch; local
325 struct i915_vma *vma[2], *batch; local
    [all...]
i915_gem_client_blt.c 166 struct i915_vma *batch; local
184 batch = intel_emit_vma_fill_blt(w->ce, vma, w->value);
185 if (IS_ERR(batch)) {
186 err = PTR_ERR(batch);
201 err = intel_emit_vma_mark_active(batch, rq);
221 batch->node.start, batch->node.size,
231 intel_emit_vma_release(w->ce, batch);
  /src/usr.bin/mail/
mime_detach.c 62 int batch; member in struct:__anon8280
71 detach_ctl.batch = value(ENAME_MIME_DETACH_BATCH) != NULL;
72 detach_ctl.ask = detach_ctl.batch ? 0 : 1;
97 if (!detach_ctl.batch) {
146 detach_ctl.batch = 1;
155 detach_ctl.batch = 1;
237 detach_ctl.batch = 0;
242 } while (!detach_ctl.batch);
  /src/external/gpl3/gcc.old/dist/libsanitizer/tsan/
tsan_dense_alloc.h 125 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); local
129 new(batch + i) T;
130 *(IndexT *)(batch + i) = i + 1 + fillpos * kL2Size;
132 *(IndexT*)(batch + kL2Size - 1) = 0;
134 map_[fillpos] = batch;
  /src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/
sanitizer_allocator_local_cache.h 169 void *res = c->batch[--c->count];
170 PREFETCH(c->batch[c->count - 1]);
184 c->batch[c->count++] = p;
212 void *batch[2 * TransferBatch::kMaxNumCached]; member in struct:SizeClassAllocator32LocalCache::PerClass
228 // id. 0 means the class size is large enough to store a batch within one
249 b->CopyToArray(c->batch);
260 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]);
261 // Failure to allocate a batch while releasing memory is non recoverable.
262 // TODO(alekseys): Figure out how to do it without allocating a new batch.
265 "transfer batch.\n", SanitizerToolName)
    [all...]
sanitizer_quarantine.h 33 void *batch[kSize]; member in struct:__sanitizer::QuarantineBatch
37 batch[0] = ptr;
38 this->size = size + sizeof(QuarantineBatch); // Account for the batch size.
41 // The total size of quarantined nodes recorded in this batch.
48 batch[count++] = ptr;
61 batch[count + i] = from->batch[i];
166 // require some tuning). It saves us merge attempt when the batch list
187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch));
189 PREFETCH(b->batch[i])
    [all...]
  /src/external/bsd/jemalloc/dist/test/unit/
batch_alloc.c 9 verify_batch_basic(tsd_t *tsd, void **ptrs, size_t batch, size_t usize,
11 for (size_t i = 0; i < batch; ++i) {
23 verify_batch_locality(tsd_t *tsd, void **ptrs, size_t batch, size_t usize,
27 * Checking batch locality when prof is on is feasible but
33 for (size_t i = 0, j = 0; i < batch; ++i, ++j) {
37 if (j == 0 && batch - i < nregs) {
54 release_batch(void **ptrs, size_t batch, size_t size) {
55 for (size_t i = 0; i < batch; ++i) {
123 size_t batch = base + (size_t)j; local
124 assert(batch < BATCH_MAX)
    [all...]
  /src/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/
sanitizer_allocator_local_cache.h 168 void *res = c->batch[--c->count];
169 PREFETCH(c->batch[c->count - 1]);
183 c->batch[c->count++] = p;
211 void *batch[2 * TransferBatch::kMaxNumCached]; member in struct:SizeClassAllocator32LocalCache::PerClass
227 // id. 0 means the class size is large enough to store a batch within one
248 b->CopyToArray(c->batch);
259 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]);
260 // Failure to allocate a batch while releasing memory is non recoverable.
261 // TODO(alekseys): Figure out how to do it without allocating a new batch.
264 "transfer batch.\n", SanitizerToolName)
    [all...]
sanitizer_quarantine.h 32 void *batch[kSize]; member in struct:__sanitizer::QuarantineBatch
36 batch[0] = ptr;
37 this->size = size + sizeof(QuarantineBatch); // Account for the batch size.
40 // The total size of quarantined nodes recorded in this batch.
47 batch[count++] = ptr;
60 batch[count + i] = from->batch[i];
166 // require some tuning). It saves us merge attempt when the batch list
187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch));
189 PREFETCH(b->batch[i])
    [all...]
  /src/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/
sanitizer_allocator_local_cache.h 168 void *res = c->batch[--c->count];
169 PREFETCH(c->batch[c->count - 1]);
183 c->batch[c->count++] = p;
211 void *batch[2 * TransferBatch::kMaxNumCached]; member in struct:SizeClassAllocator32LocalCache::PerClass
227 // id. 0 means the class size is large enough to store a batch within one
248 b->CopyToArray(c->batch);
259 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]);
260 // Failure to allocate a batch while releasing memory is non recoverable.
261 // TODO(alekseys): Figure out how to do it without allocating a new batch.
264 "transfer batch.\n", SanitizerToolName)
    [all...]
  /src/external/gpl3/gcc/dist/libsanitizer/tsan/
tsan_dense_alloc.h 113 // The stack itself uses Block::next links, while the batch within each
114 // stack node uses Block::batch links.
122 IndexT batch; member in struct:__tsan::DenseSlabAlloc::Block
131 // Pop 1 batch of nodes from the freelist.
146 idx = MapBlock(idx)->batch;
151 // Build a batch of at most Cache::kSize / 2 nodes linked by Block::batch.
156 ptr->batch = head_idx;
180 T *batch = (T *)MmapOrDie(kL2Size * sizeof(T), name_); local
181 map_[fillpos] = batch;
    [all...]
  /src/games/boggle/boggle/
bog.c 126 static int batch; variable
141 batch = debug = reuse = selfuse = sflag = 0;
149 batch = 1;
199 if (batch && bspec == NULL)
206 if (batch) {
  /src/sys/external/bsd/drm/dist/shared-core/
i915_dma.c 486 drm_i915_batchbuffer_t * batch)
489 struct drm_clip_rect __user *boxes = batch->cliprects;
490 int nbox = batch->num_cliprects;
494 if ((batch->start | batch->used) & 0x7) {
506 batch->DR1, batch->DR4);
515 OUT_RING(batch->start);
518 OUT_RING(batch->start | MI_BATCH_NON_SECURE);
524 OUT_RING(batch->start | MI_BATCH_NON_SECURE)
613 drm_i915_batchbuffer_t *batch = data; local
    [all...]
  /src/sys/kern/
kern_runq.c 1037 int batch; local
1050 batch = (l->l_rticksum > l->l_slpticksum);
1051 if (batch != 0) {
1053 batch = 0;
1063 sched_pstats_hook(l, batch);
  /src/external/bsd/openldap/dist/libraries/liblmdb/
mdb_load.c 393 int batch = 0; local
457 batch++;
458 if (batch == 100) {
481 batch = 0;
  /src/external/gpl3/gcc/dist/contrib/mdcompact/
mdcompact.el 41 ;; emacs -batch -l ./contrib/mdcompact/mdcompact.el -f mdcomp-run-directory ./gcc/config/arm/
290 (defun mdcomp-batch-run-directory ()
  /src/external/gpl3/gdb/dist/gdb/python/
py-tui.c 186 tui_batch_rendering batch; local
214 tui_batch_rendering batch; local
230 tui_batch_rendering batch; local
254 tui_batch_rendering batch; local
272 tui_batch_rendering batch; local
  /src/sys/external/bsd/drm2/dist/drm/i915/
i915_request.h 270 /** Batch buffer related to this request if any (used for
273 struct i915_vma *batch; member in struct:i915_request

Completed in 43 milliseconds

1 2