if (sz < PMD_SIZE)
                        return PAGE_SIZE;
 
-               /*
-                * The address we faulted on is backed by a transparent huge
-                * page.  However, because we map the compound huge page and
-                * not the individual tail page, we need to transfer the
-                * refcount to the head page.  We have to be careful that the
-                * THP doesn't start to split while we are adjusting the
-                * refcounts.
-                *
-                * We are sure this doesn't happen, because mmu_invalidate_retry
-                * was successful and we are holding the mmu_lock, so if this
-                * THP is trying to split, it will be blocked in the mmu
-                * notifier before touching any of the pages, specifically
-                * before being able to call __split_huge_page_refcount().
-                *
-                * We can therefore safely transfer the refcount from PG_tail
-                * to PG_head and switch the pfn from a tail page to the head
-                * page accordingly.
-                */
                *ipap &= PMD_MASK;
-               kvm_release_pfn_clean(pfn);
                pfn &= ~(PTRS_PER_PMD - 1);
-               get_page(pfn_to_page(pfn));
                *pfnp = pfn;
 
                return PMD_SIZE;