HomeSort by: relevance | last modified time | path
    Searched refs:ce (Results 1 - 25 of 100) sorted by relevancy

1 2 3 4

  /src/sys/external/bsd/drm2/dist/drm/i915/gt/
intel_context.h 22 #define CE_TRACE(ce, fmt, ...) do { \
23 const struct intel_context *ce__ = (ce); \
29 void intel_context_init(struct intel_context *ce,
31 void intel_context_fini(struct intel_context *ce);
36 int intel_context_alloc_state(struct intel_context *ce);
38 void intel_context_free(struct intel_context *ce);
42 * @ce - the context
48 static inline int intel_context_lock_pinned(struct intel_context *ce)
49 __acquires(ce->pin_mutex)
51 return mutex_lock_interruptible(&ce->pin_mutex)
    [all...]
intel_context.c 35 void intel_context_free(struct intel_context *ce)
37 kmem_cache_free(global.slab_ce, ce);
43 struct intel_context *ce; local in function:intel_context_create
45 ce = intel_context_alloc();
46 if (!ce)
49 intel_context_init(ce, engine);
50 return ce;
53 int intel_context_alloc_state(struct intel_context *ce)
57 if (mutex_lock_interruptible(&ce->pin_mutex))
60 if (!test_bit(CONTEXT_ALLOC_BIT, &ce->flags))
227 struct intel_context *ce = container_of(active, typeof(*ce), active); local in function:__intel_context_retire
243 struct intel_context *ce = container_of(active, typeof(*ce), active); local in function:__intel_context_active
    [all...]
intel_context_types.h 30 int (*alloc)(struct intel_context *ce);
32 int (*pin)(struct intel_context *ce);
33 void (*unpin)(struct intel_context *ce);
35 void (*enter)(struct intel_context *ce);
36 void (*exit)(struct intel_context *ce);
38 void (*reset)(struct intel_context *ce);
47 #define intel_context_inflight(ce) ptr_mask_bits((ce)->inflight, 2)
48 #define intel_context_inflight_count(ce) ptr_unmask_bits((ce)->inflight, 2
    [all...]
selftest_mocs.c 208 struct intel_context *ce)
218 rq = intel_context_create_request(ce);
232 if (!err && ce->engine->class == RENDER_CLASS)
244 err = check_mocs_table(ce->engine, &arg->table, &vaddr);
245 if (!err && ce->engine->class == RENDER_CLASS)
246 err = check_l3cc_table(ce->engine, &arg->table, &vaddr);
295 struct intel_context *ce; local in function:live_mocs_clean
297 ce = intel_context_create(engine);
298 if (IS_ERR(ce)) {
299 err = PTR_ERR(ce);
389 struct intel_context *ce; local in function:live_mocs_reset
    [all...]
intel_engine_pm.c 28 struct intel_context *ce; local in function:__engine_unpark
44 ce = engine->kernel_context;
45 if (ce) {
46 GEM_BUG_ON(test_bit(CONTEXT_VALID_BIT, &ce->flags));
49 if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM) && ce->state) {
50 struct drm_i915_gem_object *obj = ce->state->obj;
61 ce->ops->reset(ce);
73 static inline unsigned long __timeline_mark_lock(struct intel_context *ce)
78 mutex_acquire(&ce->timeline->mutex.dep_map, 2, 0, _THIS_IP_)
148 struct intel_context *ce = engine->kernel_context; local in function:switch_to_kernel_context
    [all...]
mock_engine.c 133 static void mock_context_unpin(struct intel_context *ce)
139 struct intel_context *ce = container_of(ref, typeof(*ce), ref); local in function:mock_context_destroy
141 GEM_BUG_ON(intel_context_is_pinned(ce));
143 if (test_bit(CONTEXT_ALLOC_BIT, &ce->flags)) {
144 mock_ring_free(ce->ring);
145 mock_timeline_unpin(ce->timeline);
148 intel_context_fini(ce);
149 intel_context_free(ce);
152 static int mock_context_alloc(struct intel_context *ce)
328 struct intel_context *ce; local in function:mock_engine_init
    [all...]
intel_breadcrumbs.c 98 check_signal_order(struct intel_context *ce, struct i915_request *rq)
100 if (!list_is_last(&rq->signal_link, &ce->signals) &&
105 if (!list_is_first(&rq->signal_link, &ce->signals) &&
160 struct intel_context *ce, *cn; local in function:signal_irq_work
169 list_for_each_entry_safe(ce, cn, &b->signalers, signal_link) {
170 GEM_BUG_ON(list_empty(&ce->signals));
172 list_for_each_safe(pos, next, &ce->signals) {
176 GEM_BUG_ON(!check_signal_order(ce, rq));
202 if (!list_is_first(pos, &ce->signals)) {
204 __list_del_many(&ce->signals, pos)
305 struct intel_context *ce = rq->context; local in function:i915_request_enable_breadcrumb
362 struct intel_context *ce = rq->context; local in function:i915_request_cancel_breadcrumb
377 struct intel_context *ce; local in function:intel_engine_print_breadcrumbs
    [all...]
selftest_rc6.c 78 static const u32 *__live_rc6_ctx(struct intel_context *ce)
85 rq = intel_context_create_request(ce);
101 *cs++ = ce->timeline->hwsp_offset + 8;
161 struct intel_context *ce; local in function:live_rc6_ctx_wa
168 ce = intel_context_create(engine);
169 if (IS_ERR(ce)) {
170 err = PTR_ERR(ce);
175 res = __live_rc6_ctx(ce);
177 intel_context_put(ce);
  /src/lib/libc/citrus/
citrus_stdenc.h 65 _citrus_stdenc_init_state(struct _citrus_stdenc * __restrict ce,
68 _DIAGASSERT(ce && ce->ce_ops && ce->ce_ops->eo_init_state);
69 return (*ce->ce_ops->eo_init_state)(ce, ps);
73 _citrus_stdenc_mbtocs(struct _citrus_stdenc * __restrict ce,
79 _DIAGASSERT(ce && ce->ce_ops && ce->ce_ops->eo_mbtocs)
    [all...]
citrus_stdenc.c 57 get_state_desc_default(struct _citrus_stdenc * __restrict ce,
72 struct _citrus_stdenc *ce; local in function:_citrus_stdenc_open
83 ce = malloc(sizeof(*ce));
84 if (ce==NULL) {
88 ce->ce_ops = NULL;
89 ce->ce_closure = NULL;
90 ce->ce_module = NULL;
91 ce->ce_traits = NULL;
97 ce->ce_module = handle
    [all...]
citrus_stdenc_template.h 64 _FUNCNAME(stdenc_init)(struct _citrus_stdenc * __restrict ce,
85 ce->ce_closure = ei;
87 et->et_mb_cur_max = _ENCODING_MB_CUR_MAX(_CE_TO_EI(ce));
93 _FUNCNAME(stdenc_uninit)(struct _citrus_stdenc * __restrict ce)
95 if (ce) {
96 _FUNCNAME(encoding_module_uninit)(_CE_TO_EI(ce));
97 free(ce->ce_closure);
102 _FUNCNAME(stdenc_init_state)(struct _citrus_stdenc * __restrict ce,
105 _FUNCNAME(init_state)(_CE_TO_EI(ce), _TO_STATE(ps));
111 _FUNCNAME(stdenc_mbtocs)(struct _citrus_stdenc * __restrict ce,
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/
i915_gem_object_blt.h 20 struct i915_vma *intel_emit_vma_fill_blt(struct intel_context *ce,
24 struct i915_vma *intel_emit_vma_copy_blt(struct intel_context *ce,
29 void intel_emit_vma_release(struct intel_context *ce, struct i915_vma *vma);
32 struct intel_context *ce,
37 struct intel_context *ce);
i915_gem_client_blt.h 18 struct intel_context *ce,
i915_gem_object_blt.c 20 struct i915_vma *intel_emit_vma_fill_blt(struct intel_context *ce,
24 struct drm_i915_private *i915 = ce->vm->i915;
35 GEM_BUG_ON(intel_engine_is_virtual(ce->engine));
36 intel_engine_pm_get(ce->engine);
41 pool = intel_engine_get_pool(ce->engine, size);
86 intel_gt_chipset_flush(ce->vm->gt);
90 batch = i915_vma_instance(pool->obj, ce->vm, NULL);
106 intel_engine_pm_put(ce->engine);
125 void intel_emit_vma_release(struct intel_context *ce, struct i915_vma *vma)
129 intel_engine_pm_put(ce->engine)
    [all...]
  /src/sys/kern/
subr_callback.c 62 callback_register(struct callback_head *ch, struct callback_entry *ce,
67 ce->ce_func = fn;
68 ce->ce_obj = obj;
70 TAILQ_INSERT_TAIL(&ch->ch_q, ce, ce_q);
76 callback_unregister(struct callback_head *ch, struct callback_entry *ce)
82 if (__predict_false(ch->ch_next == ce)) {
83 ch->ch_next = TAILQ_NEXT(ce, ce_q);
85 TAILQ_REMOVE(&ch->ch_q, ce, ce_q);
93 struct callback_entry *ce; local in function:callback_runone
99 ce = ch->ch_next
    [all...]
  /src/sys/external/bsd/drm2/dist/drm/i915/selftests/
mock_request.h 35 mock_request(struct intel_context *ce, unsigned long delay);
mock_request.c 36 mock_request(struct intel_context *ce, unsigned long delay)
41 request = intel_context_create_request(ce);
  /src/sys/external/bsd/drm2/dist/drm/nouveau/nvkm/engine/ce/
priv.h 6 #include <engine/ce.h>
nouveau_nvkm_engine_ce_gk104.c 52 gk104_ce_intr_launcherr(struct nvkm_engine *ce, const u32 base)
54 struct nvkm_subdev *subdev = &ce->subdev;
64 gk104_ce_intr(struct nvkm_engine *ce)
66 const u32 base = (ce->subdev.index - NVKM_ENGINE_CE0) * 0x1000;
67 struct nvkm_subdev *subdev = &ce->subdev;
82 gk104_ce_intr_launcherr(ce, base);
nouveau_nvkm_engine_ce_gp100.c 54 gp100_ce_intr_launcherr(struct nvkm_engine *ce, const u32 base)
56 struct nvkm_subdev *subdev = &ce->subdev;
65 gp100_ce_intr(struct nvkm_engine *ce)
67 const u32 base = (ce->subdev.index - NVKM_ENGINE_CE0) * 0x80;
68 struct nvkm_subdev *subdev = &ce->subdev;
83 gp100_ce_intr_launcherr(ce, base);
nouveau_nvkm_engine_ce_gf100.c 35 gf100_ce_init(struct nvkm_falcon *ce)
37 struct nvkm_device *device = ce->engine.subdev.device;
38 const int index = ce->engine.subdev.index - NVKM_ENGINE_CE0;
39 nvkm_wr32(device, ce->addr + 0x084, index);
  /src/usr.bin/indent/
.indent.pro 5 -ce /* Place '} else' on the same line. */
  /src/sys/external/bsd/drm2/dist/drm/i915/gem/selftests/
igt_gem_utils.c 26 struct intel_context *ce; local in function:igt_request_alloc
34 ce = i915_gem_context_get_engine(ctx, engine->legacy_idx);
35 if (IS_ERR(ce))
36 return ERR_CAST(ce);
38 rq = intel_context_create_request(ce);
39 intel_context_put(ce);
112 int igt_gpu_fill_dw(struct intel_context *ce,
121 GEM_BUG_ON(!intel_engine_can_store_dword(ce->engine));
128 rq = intel_context_create_request(ce);
135 if (INTEL_GEN(ce->vm->i915) <= 5
    [all...]
i915_gem_context.c 191 struct intel_context *ce[2]; member in struct:parallel_switch
206 for (n = 0; !err && n < ARRAY_SIZE(arg->ce); n++) {
209 rq = i915_request_create(arg->ce[n]);
232 pr_info("%s: %lu switches (sync)\n", arg->ce[0]->engine->name, count);
246 for (n = 0; n < ARRAY_SIZE(arg->ce); n++) {
250 rq = i915_request_create(arg->ce[n]);
273 pr_info("%s: %lu switches (many)\n", arg->ce[0]->engine->name, count);
290 struct intel_context *ce; local in function:live_parallel_switch
323 for_each_gem_engine(ce, engines, it) {
324 err = intel_context_pin(ce);
680 struct intel_context *ce; local in function:igt_ctx_exec
813 struct intel_context *ce; local in function:igt_shared_ctx_exec
1246 struct intel_context *ce; local in function:__igt_ctx_sseu
1356 struct intel_context *ce; local in function:igt_ctx_readonly
    [all...]
igt_gem_utils.h 30 int igt_gpu_fill_dw(struct intel_context *ce,

Completed in 20 milliseconds

1 2 3 4