]> www.infradead.org Git - users/jedix/linux-maple.git/commitdiff
mm: memory_hotplug: enumerate all supported section flags
authorMuchun Song <songmuchun@bytedance.com>
Fri, 17 Jun 2022 13:56:49 +0000 (21:56 +0800)
committerLiam R. Howlett <Liam.Howlett@oracle.com>
Wed, 20 Jul 2022 00:15:10 +0000 (20:15 -0400)
Patch series "make hugetlb_optimize_vmemmap compatible with
memmap_on_memory", v3.

This series makes hugetlb_optimize_vmemmap compatible with
memmap_on_memory.

This patch (of 2):

We are almost running out of section flags, only one bit is available in
the worst case (powerpc with 256k pages).  However, there are still some
free bits (in ->section_mem_map) on other architectures (e.g.  x86_64 has
10 bits available, arm64 has 8 bits available with worst case of 64K
pages).  We have hard coded those numbers in code, it is inconvenient to
use those bits on other architectures except powerpc.  So transfer those
section flags to enumeration to make it easy to add new section flags in
the future.  Also, move SECTION_TAINT_ZONE_DEVICE into the scope of
CONFIG_ZONE_DEVICE to save a bit on non-zone-device case.

Link: https://lkml.kernel.org/r/20220617135650.74901-1-songmuchun@bytedance.com
Link: https://lkml.kernel.org/r/20220617135650.74901-2-songmuchun@bytedance.com
Signed-off-by: Muchun Song <songmuchun@bytedance.com>
Reviewed-by: David Hildenbrand <david@redhat.com>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Mike Kravetz <mike.kravetz@oracle.com>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: Paul E. McKenney <paulmck@kernel.org>
Cc: Xiongchun Duan <duanxiongchun@bytedance.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
include/linux/mmzone.h
mm/memory_hotplug.c
mm/sparse.c

index aab70355d64f3f15cd8be7be8016b26239ac3992..932843c6459bc6b3e6315332ebdf131f4e8a3f9d 100644 (file)
@@ -1418,16 +1418,35 @@ extern size_t mem_section_usage_size(void);
  *      (equal SECTION_SIZE_BITS - PAGE_SHIFT), and the
  *      worst combination is powerpc with 256k pages,
  *      which results in PFN_SECTION_SHIFT equal 6.
- * To sum it up, at least 6 bits are available.
+ * To sum it up, at least 6 bits are available on all architectures.
+ * However, we can exceed 6 bits on some other architectures except
+ * powerpc (e.g. 15 bits are available on x86_64, 13 bits are available
+ * with the worst case of 64K pages on arm64) if we make sure the
+ * exceeded bit is not applicable to powerpc.
  */
-#define SECTION_MARKED_PRESENT         (1UL<<0)
-#define SECTION_HAS_MEM_MAP            (1UL<<1)
-#define SECTION_IS_ONLINE              (1UL<<2)
-#define SECTION_IS_EARLY               (1UL<<3)
-#define SECTION_TAINT_ZONE_DEVICE      (1UL<<4)
-#define SECTION_MAP_LAST_BIT           (1UL<<5)
-#define SECTION_MAP_MASK               (~(SECTION_MAP_LAST_BIT-1))
-#define SECTION_NID_SHIFT              6
+enum {
+       SECTION_MARKED_PRESENT_BIT,
+       SECTION_HAS_MEM_MAP_BIT,
+       SECTION_IS_ONLINE_BIT,
+       SECTION_IS_EARLY_BIT,
+#ifdef CONFIG_ZONE_DEVICE
+       SECTION_TAINT_ZONE_DEVICE_BIT,
+#endif
+       SECTION_MAP_LAST_BIT,
+};
+
+enum {
+       SECTION_MARKED_PRESENT          = BIT(SECTION_MARKED_PRESENT_BIT),
+       SECTION_HAS_MEM_MAP             = BIT(SECTION_HAS_MEM_MAP_BIT),
+       SECTION_IS_ONLINE               = BIT(SECTION_IS_ONLINE_BIT),
+       SECTION_IS_EARLY                = BIT(SECTION_IS_EARLY_BIT),
+#ifdef CONFIG_ZONE_DEVICE
+       SECTION_TAINT_ZONE_DEVICE       = BIT(SECTION_TAINT_ZONE_DEVICE_BIT),
+#endif
+};
+
+#define SECTION_MAP_MASK               (~(BIT(SECTION_MAP_LAST_BIT) - 1))
+#define SECTION_NID_SHIFT              SECTION_MAP_LAST_BIT
 
 static inline struct page *__section_mem_map_addr(struct mem_section *section)
 {
@@ -1466,12 +1485,19 @@ static inline int online_section(struct mem_section *section)
        return (section && (section->section_mem_map & SECTION_IS_ONLINE));
 }
 
+#ifdef CONFIG_ZONE_DEVICE
 static inline int online_device_section(struct mem_section *section)
 {
        unsigned long flags = SECTION_IS_ONLINE | SECTION_TAINT_ZONE_DEVICE;
 
        return section && ((section->section_mem_map & flags) == flags);
 }
+#else
+static inline int online_device_section(struct mem_section *section)
+{
+       return 0;
+}
+#endif
 
 static inline int online_section_nr(unsigned long nr)
 {
index 84990a14d51af7fb6f469ec0ed3c37d9ec5a8729..a2a6d280054ffd2aac1cd7d7d9496d054d4b1048 100644 (file)
@@ -670,12 +670,18 @@ static void __meminit resize_pgdat_range(struct pglist_data *pgdat, unsigned lon
 
 }
 
+#ifdef CONFIG_ZONE_DEVICE
 static void section_taint_zone_device(unsigned long pfn)
 {
        struct mem_section *ms = __pfn_to_section(pfn);
 
        ms->section_mem_map |= SECTION_TAINT_ZONE_DEVICE;
 }
+#else
+static inline void section_taint_zone_device(unsigned long pfn)
+{
+}
+#endif
 
 /*
  * Associate the pfn range with the given zone, initializing the memmaps
index cb3bfae640365fd58c1075a5d01b05395a265dfb..e5a8a3a0edd74329f3c34d94c92c623a891faede 100644 (file)
@@ -281,7 +281,7 @@ static unsigned long sparse_encode_mem_map(struct page *mem_map, unsigned long p
 {
        unsigned long coded_mem_map =
                (unsigned long)(mem_map - (section_nr_to_pfn(pnum)));
-       BUILD_BUG_ON(SECTION_MAP_LAST_BIT > (1UL<<PFN_SECTION_SHIFT));
+       BUILD_BUG_ON(SECTION_MAP_LAST_BIT > PFN_SECTION_SHIFT);
        BUG_ON(coded_mem_map & ~SECTION_MAP_MASK);
        return coded_mem_map;
 }