}
 
 /**
- * drm_sched_run_job_queue - enqueue run-job work
+ * __drm_sched_run_job_queue - enqueue run-job work
  * @sched: scheduler instance
  */
-static void drm_sched_run_job_queue(struct drm_gpu_scheduler *sched)
+static void __drm_sched_run_job_queue(struct drm_gpu_scheduler *sched)
 {
        if (!READ_ONCE(sched->pause_submit))
                queue_work(sched->submit_wq, &sched->work_run_job);
 void drm_sched_wakeup_if_can_queue(struct drm_gpu_scheduler *sched)
 {
        if (drm_sched_can_queue(sched))
-               drm_sched_run_job_queue(sched);
+               __drm_sched_run_job_queue(sched);
 }
 
 /**
 EXPORT_SYMBOL(drm_sched_pick_best);
 
 /**
- * drm_sched_run_job_queue_if_ready - enqueue run-job work if ready
+ * drm_sched_run_job_queue - enqueue run-job work if there are ready entities
  * @sched: scheduler instance
  */
-static void drm_sched_run_job_queue_if_ready(struct drm_gpu_scheduler *sched)
+static void drm_sched_run_job_queue(struct drm_gpu_scheduler *sched)
 {
        if (drm_sched_select_entity(sched))
-               drm_sched_run_job_queue(sched);
+               __drm_sched_run_job_queue(sched);
 }
 
 /**
                sched->ops->free_job(job);
 
        drm_sched_run_free_queue(sched);
-       drm_sched_run_job_queue_if_ready(sched);
+       drm_sched_run_job_queue(sched);
 }
 
 /**
        }
 
        wake_up(&sched->job_scheduled);
-       drm_sched_run_job_queue_if_ready(sched);
+       drm_sched_run_job_queue(sched);
 }
 
 /**