| /src/sys/external/bsd/drm2/dist/drm/i915/ |
| i915_request.h | 54 #define RQ_TRACE(rq, fmt, ...) do { \ 55 const struct i915_request *rq__ = (rq); \ 234 * path would be rq->hw_context->ring->timeline->hwsp_seqno. 313 void __i915_request_queue(struct i915_request *rq, 316 bool i915_request_retire(struct i915_request *rq); 317 void i915_request_retire_upto(struct i915_request *rq); 329 i915_request_get(struct i915_request *rq) 331 return to_request(dma_fence_get(&rq->fence)); 335 i915_request_get_rcu(struct i915_request *rq) 337 return to_request(dma_fence_get_rcu(&rq->fence)) [all...] |
| i915_request.c | 52 void (*hook)(struct i915_request *rq, struct dma_fence *signal); 112 struct i915_request *rq = to_request(fence); local 122 i915_sw_fence_fini(&rq->submit); 123 i915_sw_fence_fini(&rq->semaphore); 126 kmem_cache_free(global.slab_requests, rq); 157 static void __notify_execute_cb(struct i915_request *rq) 161 lockdep_assert_held(&rq->lock); 163 if (list_empty(&rq->execute_cb)) 166 list_for_each_entry(cb, &rq->execute_cb, link) 179 INIT_LIST_HEAD(&rq->execute_cb) 558 struct i915_request *rq, *rn; local 568 struct i915_request *rq; local 598 struct i915_request *rq = arg; local 615 struct i915_request *rq = arg; local 629 struct i915_request *rq; local 766 struct i915_request *rq; local [all...] |
| /src/external/gpl2/lvm2/dist/daemons/cmirrord/ |
| local.h | 20 int kernel_send(struct dm_ulog_request *rq);
|
| functions.h | 23 int local_resume(struct dm_ulog_request *rq); 26 int do_request(struct clog_request *rq, int server); 32 int log_get_state(struct dm_ulog_request *rq);
|
| functions.c | 537 * @rq 539 * rq->data should contain constructor string as follows: 544 * FIXME: Currently relies on caller to fill in rq->error 546 static int clog_dtr(struct dm_ulog_request *rq); 547 static int clog_ctr(struct dm_ulog_request *rq) 555 if (!rq->data_size) { 560 if (strlen(rq->data) > rq->data_size) { 562 LOG_ERROR("strlen(rq->data)[%d] != rq->data_size[%llu]" [all...] |
| cluster.c | 140 * @rq 144 int cluster_send(struct clog_request *rq) 153 if (!strncmp(entry->name.value, rq->u_rq.uuid, 160 rq->u_rq.error = -ENOENT; 168 rq->u_rq.luid = 0; 170 iov.iov_base = rq; 171 iov.iov_len = sizeof(struct clog_request) + rq->u_rq.data_size; 183 SHORT_UUID(rq->u_rq.uuid), count, 187 SHORT_UUID(rq->u_rq.uuid), count, 191 SHORT_UUID(rq->u_rq.uuid), count 439 struct clog_request *rq; local 798 struct clog_request *rq, *n; local 892 struct clog_request *rq, *n; local 946 struct clog_request *rq = msg; local 1167 struct clog_request *rq; local 1245 struct clog_request *rq, *n; local 1526 struct clog_request *rq, *n; local 1621 struct clog_request *rq; local [all...] |
| /src/external/bsd/openldap/dist/libraries/libldap/ |
| rq.c | 1 /* $NetBSD: rq.c,v 1.3 2025/09/05 21:16:21 christos Exp $ */ 23 __RCSID("$NetBSD: rq.c,v 1.3 2025/09/05 21:16:21 christos Exp $"); 46 struct runqueue_s* rq, 66 LDAP_STAILQ_INSERT_HEAD( &rq->task_list, entry, tnext ); 73 struct runqueue_s *rq, 80 LDAP_STAILQ_FOREACH( e, &rq->task_list, tnext ) { 89 struct runqueue_s* rq, 95 LDAP_STAILQ_FOREACH( e, &rq->task_list, tnext ) { 102 LDAP_STAILQ_REMOVE( &rq->task_list, entry, re_s, tnext ); 109 struct runqueue_s* rq, [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/selftests/ |
| i915_perf.c | 77 static int write_timestamp(struct i915_request *rq, int slot) 82 cs = intel_ring_begin(rq, 6); 87 if (INTEL_GEN(rq->i915) >= 8) 99 intel_ring_advance(rq, cs); 104 static ktime_t poll_status(struct i915_request *rq, int slot) 106 while (!intel_read_status_page(rq->engine, slot) && 107 !i915_request_completed(rq)) 117 struct i915_request *rq; local 140 rq = intel_engine_create_kernel_request(stream->engine); 141 if (IS_ERR(rq)) { [all...] |
| igt_spinner.c | 72 const struct i915_request *rq) 74 return hws->node.start + seqno_offset(rq->fence.context); 78 struct i915_request *rq, 84 err = i915_request_await_object(rq, vma->obj, 87 err = i915_vma_move_to_active(vma, rq, flags); 99 struct i915_request *rq = NULL; local 126 rq = intel_context_create_request(ce); 127 if (IS_ERR(rq)) { 128 err = PTR_ERR(rq); 132 err = move_to_active(vma, rq, 0) [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
| intel_breadcrumbs.c | 92 static inline bool __request_completed(const struct i915_request *rq) 94 return i915_seqno_passed(__hwsp_seqno(rq), rq->fence.seqno); 98 check_signal_order(struct intel_context *ce, struct i915_request *rq) 100 if (!list_is_last(&rq->signal_link, &ce->signals) && 101 i915_seqno_passed(rq->fence.seqno, 102 list_next_entry(rq, signal_link)->fence.seqno)) 105 if (!list_is_first(&rq->signal_link, &ce->signals) && 106 i915_seqno_passed(list_prev_entry(rq, signal_link)->fence.seqno, 107 rq->fence.seqno) 173 struct i915_request *rq = local 215 struct i915_request *rq = local 378 struct i915_request *rq; local [all...] |
| intel_engine_heartbeat.c | 44 static void idle_pulse(struct intel_engine_cs *engine, struct i915_request *rq) 47 i915_request_add_active_barriers(rq); 50 static void show_heartbeat(const struct i915_request *rq, 58 rq->sched.attr.priority); 69 struct i915_request *rq; local 71 rq = engine->heartbeat.systole; 72 if (rq && i915_request_completed(rq)) { 73 i915_request_put(rq); 85 rq->sched.attr.priority < I915_PRIORITY_BARRIER) 195 struct i915_request *rq; local 230 struct i915_request *rq; local [all...] |
| intel_engine_pool.h | 21 struct i915_request *rq) 23 return i915_active_add_request(&node->active, rq);
|
| selftest_hangcheck.c | 116 const struct i915_request *rq) 118 return hws->node.start + offset_in_page(sizeof(u32)*rq->fence.context); 122 struct i915_request *rq, 128 err = i915_request_await_object(rq, vma->obj, 131 err = i915_vma_move_to_active(vma, rq, flags); 143 struct i915_request *rq = NULL; local 191 rq = igt_request_alloc(h->ctx, engine); 192 if (IS_ERR(rq)) { 193 err = PTR_ERR(rq); 197 err = move_to_active(vma, rq, 0) 339 struct i915_request *rq; local 426 struct i915_request *rq; local 506 struct i915_request *rq; local 592 struct i915_request *rq; local 709 struct i915_request *rq[8] = {}; local 843 struct i915_request *rq = NULL; local 1032 struct i915_request *rq; local 1161 struct i915_request *rq; local 1392 struct i915_request *rq; local 1505 struct i915_request *rq; local 1591 struct i915_request *rq; local [all...] |
| intel_ring.h | 21 u32 *intel_ring_begin(struct i915_request *rq, unsigned int num_dwords); 22 int intel_ring_cacheline_align(struct i915_request *rq); 43 static inline void intel_ring_advance(struct i915_request *rq, u32 *cs) 53 GEM_BUG_ON((rq->ring->vaddr + rq->ring->emit) != cs); 82 static inline u32 intel_ring_offset(const struct i915_request *rq, void *addr) 85 u32 offset = addr - rq->ring->vaddr; 86 GEM_BUG_ON(offset > rq->ring->size); 87 return intel_ring_wrap(rq->ring, offset);
|
| selftest_lrc.c | 92 struct i915_request *rq; local 100 rq = igt_spinner_create_request(&spin, ce, MI_NOOP); 101 if (IS_ERR(rq)) { 102 err = PTR_ERR(rq); 106 i915_request_add(rq); 107 if (!igt_wait_for_spinner(&spin, rq)) { 149 struct i915_request *rq[2]; local 196 rq[0] = igt_spinner_create_request(&spin, ce[0], MI_ARB_CHECK); 197 if (IS_ERR(rq[0])) { 198 err = PTR_ERR(rq[0]) 316 struct i915_request *rq; local 435 struct i915_request *rq; local 470 struct i915_request *rq; local 518 struct i915_request *rq; local 624 struct i915_request *rq; local 700 struct i915_request *rq, *nop; local 980 struct i915_request *rq; local 1026 struct i915_request *rq; local 1122 struct i915_request *rq; local 1331 struct i915_request *rq; local 1380 struct i915_request *rq[2] = {}; local 1452 struct i915_request *rq[3] = {}; local 1542 struct i915_request *rq; local 1770 struct i915_request *rq; local 1819 struct i915_request *rq[ARRAY_SIZE(client)] = {}; local 1967 struct i915_request *rq; local 2097 struct i915_request *rq; local 2206 struct i915_request *rq = NULL; local 2321 struct i915_request *rq; local 2437 struct i915_request *rq; local 2527 struct i915_request *rq; local 2798 struct i915_request *rq; local 2815 struct i915_request *rq; local 3079 struct i915_request *rq; local 3197 struct i915_request *rq[16]; local 3426 struct i915_request *rq; local 3844 struct i915_request *rq; local 3957 struct i915_request *rq; local 3988 struct i915_request *rq; local [all...] |
| selftest_mocs.c | 25 static int request_add_sync(struct i915_request *rq, int err) 27 i915_request_get(rq); 28 i915_request_add(rq); 29 if (i915_request_wait(rq, 0, HZ / 5) < 0) 31 i915_request_put(rq); 36 static int request_add_spin(struct i915_request *rq, struct igt_spinner *spin) 40 i915_request_get(rq); 41 i915_request_add(rq); 42 if (spin && !igt_wait_for_spinner(spin, rq)) 44 i915_request_put(rq); 211 struct i915_request *rq; local 317 struct i915_request *rq; local [all...] |
| intel_ring_submission.c | 58 gen2_render_ring_flush(struct i915_request *rq, u32 mode) 70 cs = intel_ring_begin(rq, 2 + 3 * num_store_dw); 77 *cs++ = intel_gt_scratch_offset(rq->engine->gt, 83 intel_ring_advance(rq, cs); 89 gen4_render_ring_flush(struct i915_request *rq, u32 mode) 125 if (IS_G4X(rq->i915) || IS_GEN(rq->i915, 5)) 133 cs = intel_ring_begin(rq, i); 151 *cs++ = intel_gt_scratch_offset(rq->engine->gt, 161 *cs++ = intel_gt_scratch_offset(rq->engine->gt 793 struct i915_request *pos, *rq; local [all...] |
| intel_lrc.c | 291 static void mark_eio(struct i915_request *rq) 293 if (i915_request_completed(rq)) 296 GEM_BUG_ON(i915_request_signaled(rq)); 298 dma_fence_set_error(&rq->fence, -EIO); 299 i915_request_mark_complete(rq); 303 active_request(const struct intel_timeline * const tl, struct i915_request *rq) 305 struct i915_request *active = rq; 308 list_for_each_entry_continue_reverse(rq, &tl->requests, link) { 309 if (i915_request_completed(rq)) 312 active = rq; 1007 struct i915_request *rq, *rn, *active = NULL; local 1460 struct i915_request * const *port, *rq; local 1556 struct i915_request *rq = execlists->pending[n]; local 1727 struct i915_request *rq; local 1771 const struct i915_request *rq = *engine->execlists.active; local 1797 struct i915_request *rq; local 1858 struct i915_request *rq = READ_ONCE(ve->request); local 1967 struct i915_request *rq; local 2073 struct i915_request *rq, *rn; local 2601 struct i915_request *rq; member in struct:execlists_capture 3566 struct i915_request *rq; local 3680 struct i915_request *rq, *rn; local 4827 struct i915_request *rq; local 5249 struct i915_request *rq, *last; local 5307 struct i915_request *rq = READ_ONCE(ve->request); local [all...] |
| /src/external/bsd/openldap/dist/include/ |
| ldap_rq.h | 45 struct runqueue_s* rq, 55 struct runqueue_s* rq, 62 struct runqueue_s* rq, 68 struct runqueue_s* rq, 74 struct runqueue_s* rq, 80 struct runqueue_s* rq, 86 struct runqueue_s* rq, 92 struct runqueue_s* rq, 99 struct runqueue_s* rq
|
| /src/external/lgpl3/gmp/dist/mini-gmp/tests/ |
| t-mpq_muldiv_2exp.c | 54 mpq_t aq, rq, tq; local 62 mpq_init (rq); 75 mpq_mul_2exp (rq, aq, e); 76 t1 = mpz_scan1 (mpq_numref (rq), 0); 77 t2 = mpz_scan1 (mpq_denref (rq), 0); 78 mpq_neg (tq, rq); 88 dump ("nr", mpq_numref (rq)); 89 dump ("dr", mpq_denref (rq)); 93 mpq_div_2exp (rq, aq, e); 94 t1 = mpz_scan1 (mpq_numref (rq), 0) [all...] |
| /src/external/gpl3/gcc/dist/contrib/reghunt/examples/ |
| 30643.c | 3 void bar (struct s *ps, int *p, int *__restrict__ rp, int *__restrict__ rq) 12 rq[0] = 1;
|
| /src/external/gpl3/gcc.old/dist/contrib/reghunt/examples/ |
| 30643.c | 3 void bar (struct s *ps, int *p, int *__restrict__ rp, int *__restrict__ rq) 12 rq[0] = 1;
|
| /src/sys/external/bsd/drm2/dist/drm/i915/gt/uc/ |
| intel_guc_submission.c | 222 static void guc_add_request(struct intel_guc *guc, struct i915_request *rq) 224 struct intel_engine_cs *engine = rq->engine; 225 u32 ctx_desc = lower_32_bits(rq->context->lrc_desc); 226 u32 ring_tail = intel_ring_set_tail(rq->ring, rq->tail) / sizeof(u64); 229 ring_tail, rq->fence.seqno); 255 struct i915_request *rq = *out++; local 257 flush_ggtt_writes(rq->ring->vma); 258 guc_add_request(guc, rq); 264 static inline int rq_prio(const struct i915_request *rq) 319 struct i915_request *rq, *rn; local 356 struct i915_request **port, *rq; local 399 struct i915_request * const *port, *rq; local 412 struct i915_request *rq; local 437 struct i915_request *rq, *rn; local [all...] |
| /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/ |
| igt_gem_utils.c | 27 struct i915_request *rq; local 38 rq = intel_context_create_request(ce); 41 return rq; 116 struct i915_request *rq; local 128 rq = intel_context_create_request(ce); 129 if (IS_ERR(rq)) { 130 err = PTR_ERR(rq); 138 err = rq->engine->emit_bb_start(rq, 145 err = i915_request_await_object(rq, batch->obj, false) [all...] |
| /src/sys/external/bsd/drm2/dist/drm/scheduler/ |
| sched_entity.c | 69 entity->rq = NULL; 77 entity->rq = &sched_list[0]->sched_rq[entity->priority]; 100 assert_spin_locked(&entity->rq->sched->job_list_lock); 128 * drm_sched_entity_get_free_sched - Get the rq from rq_list with least load 132 * Return the pointer to the rq with least load. 137 struct drm_sched_rq *rq = NULL; local 152 rq = &entity->sched_list[i]->sched_rq[entity->priority]; 156 return rq; 181 if (!entity->rq) 184 sched = entity->rq->sched 494 struct drm_sched_rq *rq; local [all...] |