void amdgpu_amdkfd_gpuvm_init_mem_limits(void);
 void amdgpu_amdkfd_gpuvm_destroy_cb(struct amdgpu_device *adev,
                                struct amdgpu_vm *vm);
-void amdgpu_amdkfd_unreserve_memory_limit(struct amdgpu_bo *bo);
+void amdgpu_amdkfd_release_notify(struct amdgpu_bo *bo);
 void amdgpu_amdkfd_reserve_system_mem(uint64_t size);
 #else
 static inline
 }
 
 static inline
-void amdgpu_amdkfd_unreserve_memory_limit(struct amdgpu_bo *bo)
+void amdgpu_amdkfd_release_notify(struct amdgpu_bo *bo)
 {
 }
 #endif
 
        spin_unlock(&kfd_mem_limit.mem_limit_lock);
 }
 
-void amdgpu_amdkfd_unreserve_memory_limit(struct amdgpu_bo *bo)
+void amdgpu_amdkfd_release_notify(struct amdgpu_bo *bo)
 {
        struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
        u32 domain = bo->preferred_domains;
        }
 
        unreserve_mem_limit(adev, amdgpu_bo_size(bo), domain, sg);
+
+       kfree(bo->kfd_bo);
 }
 
 
        drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv);
        if (mem->dmabuf)
                dma_buf_put(mem->dmabuf);
-       drm_gem_object_put(&mem->bo->tbo.base);
        mutex_destroy(&mem->lock);
-       kfree(mem);
+
+       /* If this releases the last reference, it will end up calling
+        * amdgpu_amdkfd_release_notify and kfree the mem struct. That's why
+        * this needs to be the last call here.
+        */
+       drm_gem_object_put(&mem->bo->tbo.base);
 
        return ret;
 }
 
        abo = ttm_to_amdgpu_bo(bo);
 
        if (abo->kfd_bo)
-               amdgpu_amdkfd_unreserve_memory_limit(abo);
+               amdgpu_amdkfd_release_notify(abo);
 
        /* We only remove the fence if the resv has individualized. */
        WARN_ON_ONCE(bo->type == ttm_bo_type_kernel