__i915_vma_set_map_and_fenceable(vma);
 
        mutex_lock(&ggtt->vm.mutex);
-       list_move_tail(&vma->vm_link, &ggtt->vm.bound_list);
+       list_add_tail(&vma->vm_link, &ggtt->vm.bound_list);
        mutex_unlock(&ggtt->vm.mutex);
 
        GEM_BUG_ON(i915_gem_object_is_shrinkable(obj));
 
 
 static void ppgtt_destroy_vma(struct i915_address_space *vm)
 {
-       struct list_head *phases[] = {
-               &vm->bound_list,
-               &vm->unbound_list,
-               NULL,
-       }, **phase;
+       struct i915_vma *vma, *vn;
 
        mutex_lock(&vm->i915->drm.struct_mutex);
-       for (phase = phases; *phase; phase++) {
-               struct i915_vma *vma, *vn;
-
-               list_for_each_entry_safe(vma, vn, *phase, vm_link)
-                       i915_vma_destroy(vma);
-       }
+       list_for_each_entry_safe(vma, vn, &vm->bound_list, vm_link)
+               i915_vma_destroy(vma);
+       GEM_BUG_ON(!list_empty(&vm->bound_list));
        mutex_unlock(&vm->i915->drm.struct_mutex);
 }
 
 
        ppgtt_destroy_vma(vm);
 
-       GEM_BUG_ON(!list_empty(&vm->bound_list));
-       GEM_BUG_ON(!list_empty(&vm->unbound_list));
-
        vm->cleanup(vm);
        i915_address_space_fini(vm);
 
 
        stash_init(&vm->free_pages);
 
-       INIT_LIST_HEAD(&vm->unbound_list);
        INIT_LIST_HEAD(&vm->bound_list);
 }
 
        INIT_LIST_HEAD(&vma->obj_link);
        INIT_LIST_HEAD(&vma->closed_link);
 
-       mutex_lock(&vma->vm->mutex);
-       list_add(&vma->vm_link, &vma->vm->unbound_list);
-       mutex_unlock(&vma->vm->mutex);
-
        return vma;
 }
 
 
         */
        struct list_head bound_list;
 
-       /**
-        * List of vma that are not unbound.
-        */
-       struct list_head unbound_list;
-
        struct pagestash free_pages;
 
        /* Global GTT */
 
 
        spin_unlock(&obj->vma.lock);
 
-       mutex_lock(&vm->mutex);
-       list_add(&vma->vm_link, &vm->unbound_list);
-       mutex_unlock(&vm->mutex);
-
        return vma;
 
 err_vma:
        GEM_BUG_ON(!i915_gem_valid_gtt_space(vma, color));
 
        mutex_lock(&vma->vm->mutex);
-       list_move_tail(&vma->vm_link, &vma->vm->bound_list);
+       list_add_tail(&vma->vm_link, &vma->vm->bound_list);
        mutex_unlock(&vma->vm->mutex);
 
        if (vma->obj) {
 
        mutex_lock(&vma->vm->mutex);
        drm_mm_remove_node(&vma->node);
-       list_move_tail(&vma->vm_link, &vma->vm->unbound_list);
+       list_del(&vma->vm_link);
        mutex_unlock(&vma->vm->mutex);
 
        /*
        GEM_BUG_ON(drm_mm_node_allocated(&vma->node));
        GEM_BUG_ON(vma->fence);
 
-       mutex_lock(&vma->vm->mutex);
-       list_del(&vma->vm_link);
-       mutex_unlock(&vma->vm->mutex);
-
        if (vma->obj) {
                struct drm_i915_gem_object *obj = vma->obj;
 
 
        vma->pages = obj->mm.pages;
 
        mutex_lock(&vma->vm->mutex);
-       list_move_tail(&vma->vm_link, &vma->vm->bound_list);
+       list_add_tail(&vma->vm_link, &vma->vm->bound_list);
        mutex_unlock(&vma->vm->mutex);
 }