struct kfd_process_device *pdd;
        void *mem;
        struct kfd_dev *dev;
-       struct svm_range_list *svms = &p->svms;
        int idr_handle;
        long err;
        uint64_t offset = args->mmap_offset;
        /* Flush pending deferred work to avoid racing with deferred actions
         * from previous memory map changes (e.g. munmap).
         */
-       svm_range_list_lock_and_flush_work(svms, current->mm);
-       mutex_lock(&svms->lock);
+       svm_range_list_lock_and_flush_work(&p->svms, current->mm);
+       mutex_lock(&p->svms.lock);
        mmap_write_unlock(current->mm);
-       if (interval_tree_iter_first(&svms->objects,
+       if (interval_tree_iter_first(&p->svms.objects,
                                     args->va_addr >> PAGE_SHIFT,
                                     (args->va_addr + args->size - 1) >> PAGE_SHIFT)) {
                pr_err("Address: 0x%llx already allocated by SVM\n",
                        args->va_addr);
-               mutex_unlock(&svms->lock);
+               mutex_unlock(&p->svms.lock);
                return -EADDRINUSE;
        }
-       mutex_unlock(&svms->lock);
+       mutex_unlock(&p->svms.lock);
 #endif
        dev = kfd_device_by_id(args->gpu_id);
        if (!dev)