long hugetlb_unreserve_pages(struct inode *inode, long start, long end,
                                                long freed);
 int isolate_hugetlb(struct page *page, struct list_head *list);
-int get_hwpoison_huge_page(struct page *page, bool *hugetlb, bool unpoison);
+int get_hwpoison_hugetlb_folio(struct folio *folio, bool *hugetlb, bool unpoison);
 int get_huge_page_for_hwpoison(unsigned long pfn, int flags,
                                bool *migratable_cleared);
 void putback_active_hugepage(struct page *page);
        return -EBUSY;
 }
 
-static inline int get_hwpoison_huge_page(struct page *page, bool *hugetlb, bool unpoison)
+static inline int get_hwpoison_hugetlb_folio(struct folio *folio, bool *hugetlb, bool unpoison)
 {
        return 0;
 }
 
        return ret;
 }
 
-int get_hwpoison_huge_page(struct page *page, bool *hugetlb, bool unpoison)
+int get_hwpoison_hugetlb_folio(struct folio *folio, bool *hugetlb, bool unpoison)
 {
        int ret = 0;
 
        *hugetlb = false;
        spin_lock_irq(&hugetlb_lock);
-       if (PageHeadHuge(page)) {
+       if (folio_test_hugetlb(folio)) {
                *hugetlb = true;
-               if (HPageFreed(page))
+               if (folio_test_hugetlb_freed(folio))
                        ret = 0;
-               else if (HPageMigratable(page) || unpoison)
-                       ret = get_page_unless_zero(page);
+               else if (folio_test_hugetlb_migratable(folio) || unpoison)
+                       ret = folio_try_get(folio);
                else
                        ret = -EBUSY;
        }
 
 
 static int __get_hwpoison_page(struct page *page, unsigned long flags)
 {
-       struct page *head = compound_head(page);
+       struct folio *folio = page_folio(page);
        int ret = 0;
        bool hugetlb = false;
 
-       ret = get_hwpoison_huge_page(head, &hugetlb, false);
+       ret = get_hwpoison_hugetlb_folio(folio, &hugetlb, false);
        if (hugetlb)
                return ret;
 
        /*
-        * This check prevents from calling get_page_unless_zero() for any
-        * unsupported type of page in order to reduce the risk of unexpected
-        * races caused by taking a page refcount.
+        * This check prevents from calling folio_try_get() for any
+        * unsupported type of folio in order to reduce the risk of unexpected
+        * races caused by taking a folio refcount.
         */
-       if (!HWPoisonHandlable(head, flags))
+       if (!HWPoisonHandlable(&folio->page, flags))
                return -EBUSY;
 
-       if (get_page_unless_zero(head)) {
-               if (head == compound_head(page))
+       if (folio_try_get(folio)) {
+               if (folio == page_folio(page))
                        return 1;
 
                pr_info("%#lx cannot catch tail\n", page_to_pfn(page));
-               put_page(head);
+               folio_put(folio);
        }
 
        return 0;
 
 static int __get_unpoison_page(struct page *page)
 {
-       struct page *head = compound_head(page);
+       struct folio *folio = page_folio(page);
        int ret = 0;
        bool hugetlb = false;
 
-       ret = get_hwpoison_huge_page(head, &hugetlb, true);
+       ret = get_hwpoison_hugetlb_folio(folio, &hugetlb, true);
        if (hugetlb)
                return ret;