]> www.infradead.org Git - users/hch/dma-mapping.git/commitdiff
mm: huge_memory: add vma_thp_disabled() and thp_disabled_by_hw()
authorKefeng Wang <wangkefeng.wang@huawei.com>
Fri, 11 Oct 2024 10:24:44 +0000 (12:24 +0200)
committerAndrew Morton <akpm@linux-foundation.org>
Thu, 17 Oct 2024 07:28:10 +0000 (00:28 -0700)
Patch series "mm: don't install PMD mappings when THPs are disabled by the
hw/process/vma".

During testing, it was found that we can get PMD mappings in processes
where THP (and more precisely, PMD mappings) are supposed to be disabled.
While it works as expected for anon+shmem, the pagecache is the
problematic bit.

For s390 KVM this currently means that a VM backed by a file located on
filesystem with large folio support can crash when KVM tries accessing the
problematic page, because the readahead logic might decide to use a
PMD-sized THP and faulting it into the page tables will install a PMD
mapping, something that s390 KVM cannot tolerate.

This might also be a problem with HW that does not support PMD mappings,
but I did not try reproducing it.

Fix it by respecting the ways to disable THPs when deciding whether we can
install a PMD mapping.  khugepaged should already be taking care of not
collapsing if THPs are effectively disabled for the hw/process/vma.

This patch (of 2):

Add vma_thp_disabled() and thp_disabled_by_hw() helpers to be shared by
shmem_allowable_huge_orders() and __thp_vma_allowable_orders().

[david@redhat.com: rename to vma_thp_disabled(), split out thp_disabled_by_hw() ]
Link: https://lkml.kernel.org/r/20241011102445.934409-2-david@redhat.com
Fixes: 793917d997df ("mm/readahead: Add large folio readahead")
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Signed-off-by: David Hildenbrand <david@redhat.com>
Reported-by: Leo Fu <bfu@redhat.com>
Tested-by: Thomas Huth <thuth@redhat.com>
Reviewed-by: Ryan Roberts <ryan.roberts@arm.com>
Cc: Boqiao Fu <bfu@redhat.com>
Cc: Christian Borntraeger <borntraeger@linux.ibm.com>
Cc: Claudio Imbrenda <imbrenda@linux.ibm.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Janosch Frank <frankja@linux.ibm.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: <stable@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
include/linux/huge_mm.h
mm/huge_memory.c
mm/shmem.c

index 67d0ab3c3bbab60c6f0f526d56a3ba6835d4b939..ef5b80e48599c107a4fd92035f28d4c73c9c2600 100644 (file)
@@ -322,6 +322,24 @@ struct thpsize {
        (transparent_hugepage_flags &                                   \
         (1<<TRANSPARENT_HUGEPAGE_USE_ZERO_PAGE_FLAG))
 
+static inline bool vma_thp_disabled(struct vm_area_struct *vma,
+               unsigned long vm_flags)
+{
+       /*
+        * Explicitly disabled through madvise or prctl, or some
+        * architectures may disable THP for some mappings, for
+        * example, s390 kvm.
+        */
+       return (vm_flags & VM_NOHUGEPAGE) ||
+              test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags);
+}
+
+static inline bool thp_disabled_by_hw(void)
+{
+       /* If the hardware/firmware marked hugepage support disabled. */
+       return transparent_hugepage_flags & (1 << TRANSPARENT_HUGEPAGE_UNSUPPORTED);
+}
+
 unsigned long thp_get_unmapped_area(struct file *filp, unsigned long addr,
                unsigned long len, unsigned long pgoff, unsigned long flags);
 unsigned long thp_get_unmapped_area_vmflags(struct file *filp, unsigned long addr,
index 87b49ecc7b1e782bd4aeab10a20b9ea27bab72ed..2fb328880b509e356315dd814169b91528a57b97 100644 (file)
@@ -109,18 +109,7 @@ unsigned long __thp_vma_allowable_orders(struct vm_area_struct *vma,
        if (!vma->vm_mm)                /* vdso */
                return 0;
 
-       /*
-        * Explicitly disabled through madvise or prctl, or some
-        * architectures may disable THP for some mappings, for
-        * example, s390 kvm.
-        * */
-       if ((vm_flags & VM_NOHUGEPAGE) ||
-           test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags))
-               return 0;
-       /*
-        * If the hardware/firmware marked hugepage support disabled.
-        */
-       if (transparent_hugepage_flags & (1 << TRANSPARENT_HUGEPAGE_UNSUPPORTED))
+       if (thp_disabled_by_hw() || vma_thp_disabled(vma, vm_flags))
                return 0;
 
        /* khugepaged doesn't collapse DAX vma, but page fault is fine. */
index 4f11b55063631976af81e4221c7366b768db6690..c5adb987b23cf9ba5b8117ece2b467a434f7c0a3 100644 (file)
@@ -1664,12 +1664,7 @@ unsigned long shmem_allowable_huge_orders(struct inode *inode,
        loff_t i_size;
        int order;
 
-       if (vma && ((vm_flags & VM_NOHUGEPAGE) ||
-           test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags)))
-               return 0;
-
-       /* If the hardware/firmware marked hugepage support disabled. */
-       if (transparent_hugepage_flags & (1 << TRANSPARENT_HUGEPAGE_UNSUPPORTED))
+       if (thp_disabled_by_hw() || (vma && vma_thp_disabled(vma, vm_flags)))
                return 0;
 
        global_huge = shmem_huge_global_enabled(inode, index, write_end,