struct intel_ring_buffer *pipelined);
 int __must_check i915_gem_object_put_fence(struct drm_i915_gem_object *obj);
 
-static inline void
+static inline bool
 i915_gem_object_pin_fence(struct drm_i915_gem_object *obj)
 {
        if (obj->fence_reg != I915_FENCE_REG_NONE) {
                struct drm_i915_private *dev_priv = obj->base.dev->dev_private;
                dev_priv->fence_regs[obj->fence_reg].pin_count++;
-       }
+               return true;
+       } else
+               return false;
 }
 
 static inline void
 
        if (!obj->has_global_gtt_mapping)
                i915_gem_gtt_bind_object(obj, obj->cache_level);
 
-       if (obj->tiling_mode == I915_TILING_NONE)
-               ret = i915_gem_object_put_fence(obj);
-       else
-               ret = i915_gem_object_get_fence(obj, NULL);
+       ret = i915_gem_object_get_fence(obj, NULL);
        if (ret)
                goto unlock;
 
 }
 
 /**
- * i915_gem_object_get_fence - set up a fence reg for an object
+ * i915_gem_object_get_fence - set up fencing for an object
  * @obj: object to map through a fence reg
  * @pipelined: ring on which to queue the change, or NULL for CPU access
- * @interruptible: must we wait uninterruptibly for the register to retire?
  *
  * When mapping objects through the GTT, userspace wants to be able to write
  * to them without having to worry about swizzling if the object is tiled.
- *
  * This function walks the fence regs looking for a free one for @obj,
  * stealing one if it can't find any.
  *
  * It then sets up the reg based on the object's properties: address, pitch
  * and tiling format.
+ *
+ * For an untiled surface, this removes any existing fence.
  */
 int
 i915_gem_object_get_fence(struct drm_i915_gem_object *obj,
        struct drm_i915_fence_reg *reg;
        int ret;
 
+       if (obj->tiling_mode == I915_TILING_NONE)
+               return i915_gem_object_put_fence(obj);
+
        /* XXX disable pipelining. There are bugs. Shocking. */
        pipelined = NULL;
 
 
 
        if (has_fenced_gpu_access) {
                if (entry->flags & EXEC_OBJECT_NEEDS_FENCE) {
-                       if (obj->tiling_mode) {
-                               ret = i915_gem_object_get_fence(obj, ring);
-                               if (ret)
-                                       goto err_unpin;
+                       ret = i915_gem_object_get_fence(obj, ring);
+                       if (ret)
+                               goto err_unpin;
 
+                       if (i915_gem_object_pin_fence(obj))
                                entry->flags |= __EXEC_OBJECT_HAS_FENCE;
-                               i915_gem_object_pin_fence(obj);
-                       } else {
-                               ret = i915_gem_object_put_fence(obj);
-                               if (ret)
-                                       goto err_unpin;
-                       }
+
                        obj->pending_fenced_gpu_access = true;
                }
        }
 
         * framebuffer compression.  For simplicity, we always install
         * a fence as the cost is not that onerous.
         */
-       if (obj->tiling_mode != I915_TILING_NONE) {
-               ret = i915_gem_object_get_fence(obj, pipelined);
-               if (ret)
-                       goto err_unpin;
+       ret = i915_gem_object_get_fence(obj, pipelined);
+       if (ret)
+               goto err_unpin;
 
-               i915_gem_object_pin_fence(obj);
-       }
+       i915_gem_object_pin_fence(obj);
 
        dev_priv->mm.interruptible = true;
        return 0;