if (tile) {
                spin_lock(&drm->tile.lock);
-               tile->fence = nouveau_fence_ref((struct nouveau_fence *)fence);
+               tile->fence = (struct nouveau_fence *)fence_get(fence);
                tile->used = false;
                spin_unlock(&drm->tile.lock);
        }
        }
 
        mutex_lock_nested(&cli->mutex, SINGLE_DEPTH_NESTING);
-       ret = nouveau_fence_sync(nouveau_bo(bo), chan);
+       ret = nouveau_fence_sync(nouveau_bo(bo), chan, true);
        if (ret == 0) {
                ret = drm->ttm.move(chan, bo, &bo->mem, new_mem);
                if (ret == 0) {
 }
 
 void
-nouveau_bo_fence(struct nouveau_bo *nvbo, struct nouveau_fence *fence)
+nouveau_bo_fence(struct nouveau_bo *nvbo, struct nouveau_fence *fence, bool exclusive)
 {
        struct reservation_object *resv = nvbo->bo.resv;
 
-       reservation_object_add_excl_fence(resv, &fence->base);
+       if (exclusive)
+               reservation_object_add_excl_fence(resv, &fence->base);
+       else if (fence)
+               reservation_object_add_shared_fence(resv, &fence->base);
 }
 
 struct ttm_bo_driver nouveau_bo_driver = {
 
 void nouveau_bo_wr16(struct nouveau_bo *, unsigned index, u16 val);
 u32  nouveau_bo_rd32(struct nouveau_bo *, unsigned index);
 void nouveau_bo_wr32(struct nouveau_bo *, unsigned index, u32 val);
-void nouveau_bo_fence(struct nouveau_bo *, struct nouveau_fence *);
+void nouveau_bo_fence(struct nouveau_bo *, struct nouveau_fence *, bool exclusive);
 int  nouveau_bo_validate(struct nouveau_bo *, bool interruptible,
                         bool no_wait_gpu);
 
 
        spin_unlock_irqrestore(&dev->event_lock, flags);
 
        /* Synchronize with the old framebuffer */
-       ret = nouveau_fence_sync(old_bo, chan);
+       ret = nouveau_fence_sync(old_bo, chan, false);
        if (ret)
                goto fail;
 
                goto fail_unpin;
 
        /* synchronise rendering channel with the kernel's channel */
-       ret = nouveau_fence_sync(new_bo, chan);
+       ret = nouveau_fence_sync(new_bo, chan, false);
        if (ret) {
                ttm_bo_unreserve(&new_bo->bo);
                goto fail_unpin;
        /* Update the crtc struct and cleanup */
        crtc->primary->fb = fb;
 
-       nouveau_bo_fence(old_bo, fence);
+       nouveau_bo_fence(old_bo, fence, false);
        ttm_bo_unreserve(&old_bo->bo);
        if (old_bo != new_bo)
                nouveau_bo_unpin(old_bo);
 
 }
 
 int
-nouveau_fence_sync(struct nouveau_bo *nvbo, struct nouveau_channel *chan)
+nouveau_fence_sync(struct nouveau_bo *nvbo, struct nouveau_channel *chan, bool exclusive)
 {
        struct nouveau_fence_chan *fctx = chan->fence;
-       struct fence *fence = NULL;
+       struct fence *fence;
        struct reservation_object *resv = nvbo->bo.resv;
        struct reservation_object_list *fobj;
+       struct nouveau_fence *f;
        int ret = 0, i;
 
+       if (!exclusive) {
+               ret = reservation_object_reserve_shared(resv);
+
+               if (ret)
+                       return ret;
+       }
+
+       fobj = reservation_object_get_list(resv);
        fence = reservation_object_get_excl(resv);
 
-       if (fence && !fence_is_signaled(fence)) {
-               struct nouveau_fence *f = from_fence(fence);
-               struct nouveau_channel *prev = f->channel;
+       if (fence && (!exclusive || !fobj || !fobj->shared_count)) {
+               struct nouveau_channel *prev = NULL;
 
-               if (prev != chan) {
-                       ret = fctx->sync(f, prev, chan);
-                       if (unlikely(ret))
-                               ret = nouveau_fence_wait(f, true, true);
-               }
-       }
+               f = nouveau_local_fence(fence, chan->drm);
+               if (f)
+                       prev = f->channel;
+
+               if (!prev || (prev != chan && (ret = fctx->sync(f, prev, chan))))
+                       ret = fence_wait(fence, true);
 
-       if (ret)
                return ret;
+       }
 
-       fobj = reservation_object_get_list(resv);
-       if (!fobj)
+       if (!exclusive || !fobj)
                return ret;
 
        for (i = 0; i < fobj->shared_count && !ret; ++i) {
+               struct nouveau_channel *prev = NULL;
+
                fence = rcu_dereference_protected(fobj->shared[i],
                                                reservation_object_held(resv));
 
-               /* should always be true, for now */
-               if (!nouveau_local_fence(fence, chan->drm))
+               f = nouveau_local_fence(fence, chan->drm);
+               if (f)
+                       prev = f->channel;
+
+               if (!prev || (ret = fctx->sync(f, prev, chan)))
                        ret = fence_wait(fence, true);
+
+               if (ret)
+                       break;
        }
 
        return ret;
        *pfence = NULL;
 }
 
-struct nouveau_fence *
-nouveau_fence_ref(struct nouveau_fence *fence)
-{
-       if (fence)
-               fence_get(&fence->base);
-       return fence;
-}
-
 int
 nouveau_fence_new(struct nouveau_channel *chan, bool sysmem,
                  struct nouveau_fence **pfence)
 
 
 int  nouveau_fence_new(struct nouveau_channel *, bool sysmem,
                       struct nouveau_fence **);
-struct nouveau_fence *
-nouveau_fence_ref(struct nouveau_fence *);
 void nouveau_fence_unref(struct nouveau_fence **);
 
 int  nouveau_fence_emit(struct nouveau_fence *, struct nouveau_channel *);
 bool nouveau_fence_done(struct nouveau_fence *);
 void nouveau_fence_work(struct fence *, void (*)(void *), void *);
 int  nouveau_fence_wait(struct nouveau_fence *, bool lazy, bool intr);
-int  nouveau_fence_sync(struct nouveau_bo *, struct nouveau_channel *);
+int  nouveau_fence_sync(struct nouveau_bo *, struct nouveau_channel *, bool exclusive);
 
 struct nouveau_fence_chan {
        spinlock_t lock;
 
 nouveau_gem_object_unmap(struct nouveau_bo *nvbo, struct nouveau_vma *vma)
 {
        const bool mapped = nvbo->bo.mem.mem_type != TTM_PL_SYSTEM;
+       struct reservation_object *resv = nvbo->bo.resv;
+       struct reservation_object_list *fobj;
        struct fence *fence = NULL;
 
+       fobj = reservation_object_get_list(resv);
+
        list_del(&vma->head);
 
-       if (mapped)
+       if (fobj && fobj->shared_count > 1)
+               ttm_bo_wait(&nvbo->bo, true, false, false);
+       else if (fobj && fobj->shared_count == 1)
+               fence = rcu_dereference_protected(fobj->shared[0],
+                                               reservation_object_held(resv));
+       else
                fence = reservation_object_get_excl(nvbo->bo.resv);
 
-       if (fence) {
+       if (fence && mapped) {
                nouveau_fence_work(fence, nouveau_gem_object_delete, vma);
        } else {
                if (mapped)
 };
 
 static void
-validate_fini_no_ticket(struct validate_op *op, struct nouveau_fence *fence)
+validate_fini_no_ticket(struct validate_op *op, struct nouveau_fence *fence,
+                       struct drm_nouveau_gem_pushbuf_bo *pbbo)
 {
        struct nouveau_bo *nvbo;
+       struct drm_nouveau_gem_pushbuf_bo *b;
 
        while (!list_empty(&op->list)) {
                nvbo = list_entry(op->list.next, struct nouveau_bo, entry);
+               b = &pbbo[nvbo->pbbo_index];
 
                if (likely(fence))
-                       nouveau_bo_fence(nvbo, fence);
+                       nouveau_bo_fence(nvbo, fence, !!b->write_domains);
 
                if (unlikely(nvbo->validate_mapped)) {
                        ttm_bo_kunmap(&nvbo->kmap);
 }
 
 static void
-validate_fini(struct validate_op *op, struct nouveau_fence *fence)
+validate_fini(struct validate_op *op, struct nouveau_fence *fence,
+             struct drm_nouveau_gem_pushbuf_bo *pbbo)
 {
-       validate_fini_no_ticket(op, fence);
+       validate_fini_no_ticket(op, fence, pbbo);
        ww_acquire_fini(&op->ticket);
 }
 
                        list_splice_tail_init(&vram_list, &op->list);
                        list_splice_tail_init(&gart_list, &op->list);
                        list_splice_tail_init(&both_list, &op->list);
-                       validate_fini_no_ticket(op, NULL);
+                       validate_fini_no_ticket(op, NULL, NULL);
                        if (unlikely(ret == -EDEADLK)) {
                                ret = ttm_bo_reserve_slowpath(&nvbo->bo, true,
                                                              &op->ticket);
        list_splice_tail(&gart_list, &op->list);
        list_splice_tail(&both_list, &op->list);
        if (ret)
-               validate_fini(op, NULL);
+               validate_fini(op, NULL, NULL);
        return ret;
 
 }
                        return ret;
                }
 
-               ret = nouveau_fence_sync(nvbo, chan);
+               ret = nouveau_fence_sync(nvbo, chan, !!b->write_domains);
                if (unlikely(ret)) {
                        if (ret != -ERESTARTSYS)
                                NV_PRINTK(error, cli, "fail post-validate sync\n");
        if (unlikely(ret < 0)) {
                if (ret != -ERESTARTSYS)
                        NV_PRINTK(error, cli, "validating bo list\n");
-               validate_fini(op, NULL);
+               validate_fini(op, NULL, NULL);
                return ret;
        }
        *apply_relocs = ret;
                                data |= r->vor;
                }
 
-               ret = ttm_bo_wait(&nvbo->bo, false, false, false);
+               ret = ttm_bo_wait(&nvbo->bo, true, false, false);
                if (ret) {
                        NV_PRINTK(error, cli, "reloc wait_idle failed: %d\n", ret);
                        break;
        }
 
 out:
-       validate_fini(&op, fence);
+       validate_fini(&op, fence, bo);
        nouveau_fence_unref(&fence);
 
 out_prevalid: