static void i915_gem_park(struct drm_i915_private *i915)
 {
-       lockdep_assert_held(&i915->drm.struct_mutex);
+       cancel_delayed_work(&i915->gem.retire_work);
 
        i915_vma_parked(i915);
 
        i915_globals_park();
 }
 
-static void idle_work_handler(struct work_struct *work)
-{
-       struct drm_i915_private *i915 =
-               container_of(work, typeof(*i915), gem.idle_work);
-       bool park;
-
-       cancel_delayed_work_sync(&i915->gem.retire_work);
-       mutex_lock(&i915->drm.struct_mutex);
-
-       intel_wakeref_lock(&i915->gt.wakeref);
-       park = (!intel_wakeref_is_active(&i915->gt.wakeref) &&
-               !work_pending(work));
-       intel_wakeref_unlock(&i915->gt.wakeref);
-       if (park)
-               i915_gem_park(i915);
-       else
-               queue_delayed_work(i915->wq,
-                                  &i915->gem.retire_work,
-                                  round_jiffies_up_relative(HZ));
-
-       mutex_unlock(&i915->drm.struct_mutex);
-}
-
 static void retire_work_handler(struct work_struct *work)
 {
        struct drm_i915_private *i915 =
                break;
 
        case INTEL_GT_PARK:
-               queue_work(i915->wq, &i915->gem.idle_work);
+               i915_gem_park(i915);
                break;
        }
 
 
 void i915_gem_init__pm(struct drm_i915_private *i915)
 {
-       INIT_WORK(&i915->gem.idle_work, idle_work_handler);
        INIT_DELAYED_WORK(&i915->gem.retire_work, retire_work_handler);
 
        i915->gem.pm_notifier.notifier_call = pm_notifier;
 
 static void disable_retire_worker(struct drm_i915_private *i915)
 {
        i915_gem_driver_unregister__shrinker(i915);
-
        intel_gt_pm_get(&i915->gt);
-
        cancel_delayed_work_sync(&i915->gem.retire_work);
-       flush_work(&i915->gem.idle_work);
 }
 
 static void restore_retire_worker(struct drm_i915_private *i915)
 
                i915_gem_shrink_all(i915);
        fs_reclaim_release(GFP_KERNEL);
 
-       if (val & DROP_IDLE) {
-               flush_delayed_work(&i915->gem.retire_work);
-               flush_work(&i915->gem.idle_work);
-       }
-
        if (val & DROP_FREED)
                i915_gem_drain_freed_objects(i915);
 
 
                 * fires, go retire requests.
                 */
                struct delayed_work retire_work;
-
-               /**
-                * When we detect an idle GPU, we want to turn on
-                * powersaving features. So once we see that there
-                * are no more requests outstanding and no more
-                * arrive within a small period of time, we fire
-                * off the idle_work.
-                */
-               struct work_struct idle_work;
        } gem;
 
        /* For i945gm vblank irq vs. C3 workaround */
 
 
        mock_device_flush(i915);
 
-       flush_work(&i915->gem.idle_work);
        i915_gem_drain_workqueue(i915);
 
        mutex_lock(&i915->drm.struct_mutex);
 {
 }
 
-static void mock_idle_work_handler(struct work_struct *work)
-{
-}
-
 static int pm_domain_resume(struct device *dev)
 {
        return pm_generic_runtime_resume(dev);
        mock_init_contexts(i915);
 
        INIT_DELAYED_WORK(&i915->gem.retire_work, mock_retire_work_handler);
-       INIT_WORK(&i915->gem.idle_work, mock_idle_work_handler);
 
        intel_timelines_init(i915);