Lines Matching defs:sched
49 * @num_sched_list: number of drm sched in sched_list
100 assert_spin_locked(&entity->rq->sched->job_list_lock);
142 struct drm_gpu_scheduler *sched = entity->sched_list[i];
145 DRM_WARN("sched%s is not ready, skipping\n", sched->name);
149 num_score = atomic_read(&sched->score);
173 struct drm_gpu_scheduler *sched;
184 sched = entity->rq->sched;
186 spin_lock(&sched->job_list_lock);
187 DRM_SPIN_WAIT_NOINTR_UNTIL(ret, &sched->job_scheduled,
188 &sched->job_list_lock,
190 spin_unlock(&sched->job_list_lock);
199 sched->job_scheduled,
203 wait_event_killable(sched->job_scheduled,
245 job->sched->ops->free_job(job);
298 struct drm_gpu_scheduler *sched = NULL;
301 sched = entity->rq->sched;
311 if (sched) {
374 spin_lock(&entity->rq->sched->job_list_lock);
375 drm_sched_wakeup(entity->rq->sched);
376 spin_unlock(&entity->rq->sched->job_list_lock);
406 struct drm_gpu_scheduler *sched = entity->rq->sched;
422 if (s_fence && s_fence->sched == sched) {
457 struct drm_gpu_scheduler *sched = entity->rq->sched;
465 sched->ops->dependency(sched_job, entity))) {
531 atomic_inc(&entity->rq->sched->score);
551 spin_lock(&entity->rq->sched->job_list_lock);
552 drm_sched_wakeup(entity->rq->sched);
553 spin_unlock(&entity->rq->sched->job_list_lock);