/src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
amdgpu_job.c | 36 static void amdgpu_job_timedout(struct drm_sched_job *s_job) 38 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); 39 struct amdgpu_job *job = to_amdgpu_job(s_job); 44 if (amdgpu_ring_soft_recovery(ring, job->vmid, s_job->s_fence->parent)) { 46 s_job->sched->name); 123 static void amdgpu_job_free_cb(struct drm_sched_job *s_job) 125 struct amdgpu_ring *ring = to_amdgpu_ring(s_job->sched); 126 struct amdgpu_job *job = to_amdgpu_job(s_job); 128 drm_sched_job_cleanup(s_job); 130 amdgpu_ring_priority_put(ring, s_job->s_priority) 259 struct drm_sched_job *s_job; local in function:amdgpu_job_stop_all_jobs_on_sched [all...] |
amdgpu_debugfs.c | 1103 struct drm_sched_job *s_job; local in function:amdgpu_ib_preempt_job_recovery 1107 list_for_each_entry(s_job, &sched->ring_mirror_list, node) { 1108 fence = sched->ops->run_job(s_job); 1117 struct drm_sched_job *s_job; local in function:amdgpu_ib_preempt_mark_partial_job 1135 list_for_each_entry(s_job, &sched->ring_mirror_list, node) { 1136 job = to_amdgpu_job(s_job);
|
/src/sys/external/bsd/drm2/dist/drm/scheduler/ |
sched_main.c | 285 static void drm_sched_job_begin(struct drm_sched_job *s_job) 287 struct drm_gpu_scheduler *sched = s_job->sched; 291 list_add_tail(&s_job->node, &sched->ring_mirror_list); 396 struct drm_sched_job *s_job, *tmp; local in function:drm_sched_stop 421 list_for_each_entry_safe_reverse(s_job, tmp, &sched->ring_mirror_list, node) { 422 if (s_job->s_fence->parent && 423 dma_fence_remove_callback(s_job->s_fence->parent, 424 &s_job->cb)) { 432 list_del_init(&s_job->node); 436 * Wait for job's HW fence callback to finish using s_job 475 struct drm_sched_job *s_job, *tmp; local in function:drm_sched_start 522 struct drm_sched_job *s_job, *tmp; local in function:drm_sched_resubmit_jobs 668 struct drm_sched_job *s_job = container_of(cb, struct drm_sched_job, cb); local in function:drm_sched_process_job [all...] |