static int kvm_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, bool write)
 {
        int ret = 0;
-       kvm_pfn_t pfn = 0;
        kvm_pte_t *ptep, changed, new;
        gfn_t gfn = gpa >> PAGE_SHIFT;
        struct kvm *kvm = vcpu->kvm;
        struct kvm_memory_slot *slot;
-       struct page *page;
 
        spin_lock(&kvm->mmu_lock);
 
 
        /* Track access to pages marked old */
        new = kvm_pte_mkyoung(*ptep);
-       /* call kvm_set_pfn_accessed() after unlock */
-
        if (write && !kvm_pte_dirty(new)) {
                if (!kvm_pte_write(new)) {
                        ret = -EFAULT;
        }
 
        changed = new ^ (*ptep);
-       if (changed) {
+       if (changed)
                kvm_set_pte(ptep, new);
-               pfn = kvm_pte_pfn(new);
-               page = kvm_pfn_to_refcounted_page(pfn);
-               if (page)
-                       get_page(page);
-       }
-       spin_unlock(&kvm->mmu_lock);
 
-       if (changed) {
-               if (kvm_pte_young(changed))
-                       kvm_set_pfn_accessed(pfn);
-
-               if (page)
-                       put_page(page);
-       }
+       spin_unlock(&kvm->mmu_lock);
 
        if (kvm_pte_dirty(changed))
                mark_page_dirty(kvm, gfn);