struct dma_fence *fence = NULL;
        struct dma_fence **fences = NULL;
        struct dma_fence_array *cf = NULL;
-       int cur_fence = 0, i;
+       int cur_fence = 0;
        int number_tiles = hweight8(vma->tile_present);
        int err;
        u8 id;
 
        fence = cf ? &cf->base : !fence ?
                xe_exec_queue_last_fence_get(wait_exec_queue, vm) : fence;
-       if (last_op) {
-               for (i = 0; i < num_syncs; i++)
-                       xe_sync_entry_signal(&syncs[i], fence);
-       }
 
        return fence;
 
        struct dma_fence **fences = NULL;
        struct dma_fence_array *cf = NULL;
        struct xe_vm *vm = xe_vma_vm(vma);
-       int cur_fence = 0, i;
+       int cur_fence = 0;
        int number_tiles = hweight8(tile_mask);
        int err;
        u8 id;
                }
        }
 
-       if (last_op) {
-               for (i = 0; i < num_syncs; i++)
-                       xe_sync_entry_signal(&syncs[i],
-                                            cf ? &cf->base : fence);
-       }
-
        return cf ? &cf->base : fence;
 
 err_fences:
                if (IS_ERR(fence))
                        return fence;
        } else {
-               int i;
-
                xe_assert(vm->xe, xe_vm_in_fault_mode(vm));
 
                fence = xe_exec_queue_last_fence_get(wait_exec_queue, vm);
-               if (last_op) {
-                       for (i = 0; i < num_syncs; i++)
-                               xe_sync_entry_signal(&syncs[i], fence);
-               }
        }
 
-       if (last_op)
-               xe_exec_queue_last_fence_set(wait_exec_queue, vm, fence);
-
        return fence;
 }
 
             u32 num_syncs, bool first_op, bool last_op)
 {
        struct dma_fence *fence;
-       struct xe_exec_queue *wait_exec_queue = to_wait_exec_queue(vm, q);
 
        xe_vm_assert_held(vm);
        xe_bo_assert_held(xe_vma_bo(vma));
        if (IS_ERR(fence))
                return fence;
 
-       xe_vma_destroy(vma, fence);
-       if (last_op)
-               xe_exec_queue_last_fence_set(wait_exec_queue, vm, fence);
-
        return fence;
 }
 
                return xe_vm_bind(vm, vma, q, xe_vma_bo(vma), syncs, num_syncs,
                                  vma->tile_mask, true, first_op, last_op);
        } else {
-               struct dma_fence *fence =
-                       xe_exec_queue_last_fence_get(wait_exec_queue, vm);
-               int i;
-
-               /* Nothing to do, signal fences now */
-               if (last_op) {
-                       for (i = 0; i < num_syncs; i++)
-                               xe_sync_entry_signal(&syncs[i], fence);
-               }
-
-               return fence;
+               return xe_exec_queue_last_fence_get(wait_exec_queue, vm);
        }
 }
 
        return fence;
 }
 
+static void vm_bind_ioctl_ops_fini(struct xe_vm *vm, struct xe_vma_ops *vops,
+                                  struct dma_fence *fence)
+{
+       struct xe_exec_queue *wait_exec_queue = to_wait_exec_queue(vm, vops->q);
+       struct xe_vma_op *op;
+       int i;
+
+       list_for_each_entry(op, &vops->list, link) {
+               if (op->base.op == DRM_GPUVA_OP_UNMAP)
+                       xe_vma_destroy(gpuva_to_vma(op->base.unmap.va), fence);
+               else if (op->base.op == DRM_GPUVA_OP_REMAP)
+                       xe_vma_destroy(gpuva_to_vma(op->base.remap.unmap->va),
+                                      fence);
+       }
+       for (i = 0; i < vops->num_syncs; i++)
+               xe_sync_entry_signal(vops->syncs + i, fence);
+       xe_exec_queue_last_fence_set(wait_exec_queue, vm, fence);
+       dma_fence_put(fence);
+}
+
 static int vm_bind_ioctl_ops_execute(struct xe_vm *vm,
                                     struct xe_vma_ops *vops)
 {
                        xe_vm_kill(vm, false);
                        goto unlock;
                } else {
-                       dma_fence_put(fence);
+                       vm_bind_ioctl_ops_fini(vm, vops, fence);
                }
        }