struct page *page = NULL;
        spinlock_t *ptl;
        pte_t *pte, entry;
-
-       /*
-        * FOLL_PIN is not supported for follow_page(). Ordinary GUP goes via
-        * follow_hugetlb_page().
-        */
-       if (WARN_ON_ONCE(flags & FOLL_PIN))
-               return NULL;
+       int ret;
 
        hugetlb_vma_lock_read(vma);
        pte = hugetlb_walk(vma, haddr, huge_page_size(h));
        ptl = huge_pte_lock(h, mm, pte);
        entry = huge_ptep_get(pte);
        if (pte_present(entry)) {
-               page = pte_page(entry) +
-                               ((address & ~huge_page_mask(h)) >> PAGE_SHIFT);
+               page = pte_page(entry);
+
+               if (!huge_pte_write(entry)) {
+                       if (flags & FOLL_WRITE) {
+                               page = NULL;
+                               goto out;
+                       }
+
+                       if (gup_must_unshare(vma, flags, page)) {
+                               /* Tell the caller to do unsharing */
+                               page = ERR_PTR(-EMLINK);
+                               goto out;
+                       }
+               }
+
+               page += ((address & ~huge_page_mask(h)) >> PAGE_SHIFT);
+
                /*
                 * Note that page may be a sub-page, and with vmemmap
                 * optimizations the page struct may be read only.
                 * try_grab_page() should always be able to get the page here,
                 * because we hold the ptl lock and have verified pte_present().
                 */
-               if (try_grab_page(page, flags)) {
-                       page = NULL;
+               ret = try_grab_page(page, flags);
+
+               if (WARN_ON_ONCE(ret)) {
+                       page = ERR_PTR(ret);
                        goto out;
                }
        }