| /src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
| igt_spinner.h | 25 u32 *batch; member in struct:igt_spinner
|
| igt_spinner.c | 52 spin->batch = vaddr; 102 u32 *batch; local 140 batch = spin->batch; 143 *batch++ = MI_STORE_DWORD_IMM_GEN4; 144 *batch++ = lower_32_bits(hws_address(hws, rq)); 145 *batch++ = upper_32_bits(hws_address(hws, rq)); 147 *batch++ = MI_STORE_DWORD_IMM_GEN4; 148 *batch++ = 0; 149 *batch++ = hws_address(hws, rq) [all...] |
| i915_request.c | 537 * (individually), and wait for the batch to complete. We can check 654 struct i915_vma *batch) 664 batch->node.start, 665 batch->node.size, 681 struct i915_vma *batch; local 686 * (individually), and wait for the batch to complete. We can check 690 batch = empty_batch(i915); 691 if (IS_ERR(batch)) 692 return PTR_ERR(batch); 707 request = empty_request(engine, batch); 831 struct i915_vma *batch; local 965 struct i915_vma *batch; local [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
| igt_gem_utils.c | 117 struct i915_vma *batch; local 124 batch = igt_emit_store_dw(vma, offset, count, val); 125 if (IS_ERR(batch)) 126 return PTR_ERR(batch); 139 batch->node.start, batch->node.size, 144 i915_vma_lock(batch); 145 err = i915_request_await_object(rq, batch->obj, false); 147 err = i915_vma_move_to_active(batch, rq, 0); 148 i915_vma_unlock(batch); [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
| intel_renderstate.h | 37 const u32 *batch; member in struct:intel_renderstate_rodata 44 .batch = gen ## _g ## _null_state_batch, \
|
| /src/sys/external/bsd/drm2/include/linux/ |
| shrinker.h | 49 size_t batch; member in struct:shrinker
|
| /src/sys/external/bsd/compiler_rt/dist/lib/tsan/rtl/ |
| tsan_dense_alloc.h | 112 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); local 116 new(batch + i) T; 117 *(IndexT*)(batch + i) = i + 1 + fillpos_ * kL2Size; 119 *(IndexT*)(batch + kL2Size - 1) = 0; 121 map_[fillpos_++] = batch;
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
| i915_gem_object_blt.c | 27 struct i915_vma *batch; local 90 batch = i915_vma_instance(pool->obj, ce->vm, NULL); 91 if (IS_ERR(batch)) { 92 err = PTR_ERR(batch); 96 err = i915_vma_pin(batch, 0, 0, PIN_USER); 100 batch->private = pool; 101 return batch; 137 struct i915_vma *batch; local 155 batch = intel_emit_vma_fill_blt(ce, vma, value); 156 if (IS_ERR(batch)) { 211 struct i915_vma *batch; local 325 struct i915_vma *vma[2], *batch; local [all...] |
| i915_gem_client_blt.c | 166 struct i915_vma *batch; local 184 batch = intel_emit_vma_fill_blt(w->ce, vma, w->value); 185 if (IS_ERR(batch)) { 186 err = PTR_ERR(batch); 201 err = intel_emit_vma_mark_active(batch, rq); 221 batch->node.start, batch->node.size, 231 intel_emit_vma_release(w->ce, batch);
|
| /src/usr.bin/mail/ |
| mime_detach.c | 62 int batch; member in struct:__anon8280 71 detach_ctl.batch = value(ENAME_MIME_DETACH_BATCH) != NULL; 72 detach_ctl.ask = detach_ctl.batch ? 0 : 1; 97 if (!detach_ctl.batch) { 146 detach_ctl.batch = 1; 155 detach_ctl.batch = 1; 237 detach_ctl.batch = 0; 242 } while (!detach_ctl.batch);
|
| /src/external/gpl3/gcc.old/dist/libsanitizer/tsan/ |
| tsan_dense_alloc.h | 125 T *batch = (T*)MmapOrDie(kL2Size * sizeof(T), name_); local 129 new(batch + i) T; 130 *(IndexT *)(batch + i) = i + 1 + fillpos * kL2Size; 132 *(IndexT*)(batch + kL2Size - 1) = 0; 134 map_[fillpos] = batch;
|
| /src/sys/external/bsd/compiler_rt/dist/lib/sanitizer_common/ |
| sanitizer_allocator_local_cache.h | 169 void *res = c->batch[--c->count]; 170 PREFETCH(c->batch[c->count - 1]); 184 c->batch[c->count++] = p; 212 void *batch[2 * TransferBatch::kMaxNumCached]; member in struct:SizeClassAllocator32LocalCache::PerClass 228 // id. 0 means the class size is large enough to store a batch within one 249 b->CopyToArray(c->batch); 260 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); 261 // Failure to allocate a batch while releasing memory is non recoverable. 262 // TODO(alekseys): Figure out how to do it without allocating a new batch. 265 "transfer batch.\n", SanitizerToolName) [all...] |
| sanitizer_quarantine.h | 33 void *batch[kSize]; member in struct:__sanitizer::QuarantineBatch 37 batch[0] = ptr; 38 this->size = size + sizeof(QuarantineBatch); // Account for the batch size. 41 // The total size of quarantined nodes recorded in this batch. 48 batch[count++] = ptr; 61 batch[count + i] = from->batch[i]; 166 // require some tuning). It saves us merge attempt when the batch list 187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); 189 PREFETCH(b->batch[i]) [all...] |
| /src/external/bsd/jemalloc/dist/test/unit/ |
| batch_alloc.c | 9 verify_batch_basic(tsd_t *tsd, void **ptrs, size_t batch, size_t usize, 11 for (size_t i = 0; i < batch; ++i) { 23 verify_batch_locality(tsd_t *tsd, void **ptrs, size_t batch, size_t usize, 27 * Checking batch locality when prof is on is feasible but 33 for (size_t i = 0, j = 0; i < batch; ++i, ++j) { 37 if (j == 0 && batch - i < nregs) { 54 release_batch(void **ptrs, size_t batch, size_t size) { 55 for (size_t i = 0; i < batch; ++i) { 123 size_t batch = base + (size_t)j; local 124 assert(batch < BATCH_MAX) [all...] |
| /src/external/gpl3/gcc.old/dist/libsanitizer/sanitizer_common/ |
| sanitizer_allocator_local_cache.h | 168 void *res = c->batch[--c->count]; 169 PREFETCH(c->batch[c->count - 1]); 183 c->batch[c->count++] = p; 211 void *batch[2 * TransferBatch::kMaxNumCached]; member in struct:SizeClassAllocator32LocalCache::PerClass 227 // id. 0 means the class size is large enough to store a batch within one 248 b->CopyToArray(c->batch); 259 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); 260 // Failure to allocate a batch while releasing memory is non recoverable. 261 // TODO(alekseys): Figure out how to do it without allocating a new batch. 264 "transfer batch.\n", SanitizerToolName) [all...] |
| sanitizer_quarantine.h | 32 void *batch[kSize]; member in struct:__sanitizer::QuarantineBatch 36 batch[0] = ptr; 37 this->size = size + sizeof(QuarantineBatch); // Account for the batch size. 40 // The total size of quarantined nodes recorded in this batch. 47 batch[count++] = ptr; 60 batch[count + i] = from->batch[i]; 166 // require some tuning). It saves us merge attempt when the batch list 187 CHECK(kPrefetch <= ARRAY_SIZE(b->batch)); 189 PREFETCH(b->batch[i]) [all...] |
| /src/external/gpl3/gcc/dist/libsanitizer/sanitizer_common/ |
| sanitizer_allocator_local_cache.h | 168 void *res = c->batch[--c->count]; 169 PREFETCH(c->batch[c->count - 1]); 183 c->batch[c->count++] = p; 211 void *batch[2 * TransferBatch::kMaxNumCached]; member in struct:SizeClassAllocator32LocalCache::PerClass 227 // id. 0 means the class size is large enough to store a batch within one 248 b->CopyToArray(c->batch); 259 class_id, allocator, (TransferBatch *)c->batch[first_idx_to_drain]); 260 // Failure to allocate a batch while releasing memory is non recoverable. 261 // TODO(alekseys): Figure out how to do it without allocating a new batch. 264 "transfer batch.\n", SanitizerToolName) [all...] |
| /src/external/gpl3/gcc/dist/libsanitizer/tsan/ |
| tsan_dense_alloc.h | 113 // The stack itself uses Block::next links, while the batch within each 114 // stack node uses Block::batch links. 122 IndexT batch; member in struct:__tsan::DenseSlabAlloc::Block 131 // Pop 1 batch of nodes from the freelist. 146 idx = MapBlock(idx)->batch; 151 // Build a batch of at most Cache::kSize / 2 nodes linked by Block::batch. 156 ptr->batch = head_idx; 180 T *batch = (T *)MmapOrDie(kL2Size * sizeof(T), name_); local 181 map_[fillpos] = batch; [all...] |
| /src/games/boggle/boggle/ |
| bog.c | 126 static int batch; variable 141 batch = debug = reuse = selfuse = sflag = 0; 149 batch = 1; 199 if (batch && bspec == NULL) 206 if (batch) {
|
| /src/sys/external/bsd/drm/dist/shared-core/ |
| i915_dma.c | 486 drm_i915_batchbuffer_t * batch) 489 struct drm_clip_rect __user *boxes = batch->cliprects; 490 int nbox = batch->num_cliprects; 494 if ((batch->start | batch->used) & 0x7) { 506 batch->DR1, batch->DR4); 515 OUT_RING(batch->start); 518 OUT_RING(batch->start | MI_BATCH_NON_SECURE); 524 OUT_RING(batch->start | MI_BATCH_NON_SECURE) 613 drm_i915_batchbuffer_t *batch = data; local [all...] |
| /src/sys/kern/ |
| kern_runq.c | 1037 int batch; local 1050 batch = (l->l_rticksum > l->l_slpticksum); 1051 if (batch != 0) { 1053 batch = 0; 1063 sched_pstats_hook(l, batch);
|
| /src/external/bsd/openldap/dist/libraries/liblmdb/ |
| mdb_load.c | 393 int batch = 0; local 457 batch++; 458 if (batch == 100) { 481 batch = 0;
|
| /src/external/gpl3/gcc/dist/contrib/mdcompact/ |
| mdcompact.el | 41 ;; emacs -batch -l ./contrib/mdcompact/mdcompact.el -f mdcomp-run-directory ./gcc/config/arm/ 290 (defun mdcomp-batch-run-directory ()
|
| /src/external/gpl3/gdb/dist/gdb/python/ |
| py-tui.c | 186 tui_batch_rendering batch; local 214 tui_batch_rendering batch; local 230 tui_batch_rendering batch; local 254 tui_batch_rendering batch; local 272 tui_batch_rendering batch; local
|
| /src/sys/external/bsd/drm2/dist/drm/i915/ |
| i915_request.h | 270 /** Batch buffer related to this request if any (used for 273 struct i915_vma *batch; member in struct:i915_request
|