pgoff_t num_prefault)
 {
        struct vm_area_struct *vma = vmf->vma;
-       struct vm_area_struct cvma = *vma;
        struct ttm_buffer_object *bo = vma->vm_private_data;
        struct ttm_bo_device *bdev = bo->bdev;
        unsigned long page_offset;
                goto out_io_unlock;
        }
 
-       cvma.vm_page_prot = ttm_io_prot(bo->mem.placement, prot);
+       prot = ttm_io_prot(bo->mem.placement, prot);
        if (!bo->mem.bus.is_iomem) {
                struct ttm_operation_ctx ctx = {
                        .interruptible = false,
                }
        } else {
                /* Iomem should not be marked encrypted */
-               cvma.vm_page_prot = pgprot_decrypted(cvma.vm_page_prot);
+               prot = pgprot_decrypted(prot);
        }
 
        /*
                        pfn = page_to_pfn(page);
                }
 
+               /*
+                * Note that the value of @prot at this point may differ from
+                * the value of @vma->vm_page_prot in the caching- and
+                * encryption bits. This is because the exact location of the
+                * data may not be known at mmap() time and may also change
+                * at arbitrary times while the data is mmap'ed.
+                * See vmf_insert_mixed_prot() for a discussion.
+                */
                if (vma->vm_flags & VM_MIXEDMAP)
-                       ret = vmf_insert_mixed(&cvma, address,
-                                       __pfn_to_pfn_t(pfn, PFN_DEV));
+                       ret = vmf_insert_mixed_prot(vma, address,
+                                                   __pfn_to_pfn_t(pfn, PFN_DEV),
+                                                   prot);
                else
-                       ret = vmf_insert_pfn(&cvma, address, pfn);
+                       ret = vmf_insert_pfn_prot(vma, address, pfn, prot);
 
                /* Never error on prefaulted PTEs */
                if (unlikely((ret & VM_FAULT_ERROR))) {
        if (ret)
                return ret;
 
-       prot = vm_get_page_prot(vma->vm_flags);
+       prot = vma->vm_page_prot;
        ret = ttm_bo_vm_fault_reserved(vmf, prot, TTM_BO_VM_NUM_PREFAULT);
        if (ret == VM_FAULT_RETRY && !(vmf->flags & FAULT_FLAG_RETRY_NOWAIT))
                return ret;