From: Michal Hocko Date: Tue, 23 Aug 2022 09:22:30 +0000 (+0200) Subject: mm: reduce noise in show_mem for lowmem allocations X-Git-Url: https://www.infradead.org/git/?a=commitdiff_plain;h=9ea9abc5cd7ed674c548915f8340c1cf706b396c;p=users%2Fjedix%2Flinux-maple.git mm: reduce noise in show_mem for lowmem allocations While discussing early DMA pool pre-allocation failure with Christoph [1] I have realized that the allocation failure warning is rather noisy for constrained allocations like GFP_DMA{32}. Those zones are usually not populated on all nodes very often as their memory ranges are constrained. This is an attempt to reduce the ballast that doesn't provide any relevant information for those allocation failures investigation. Please note that I have only compile tested it (in my default config setup) and I am throwing it mostly to see what people think about it. [1] http://lkml.kernel.org/r/20220817060647.1032426-1-hch@lst.de Link: https://lkml.kernel.org/r/YwScVmVofIZkopkF@dhcp22.suse.cz Signed-off-by: Michal Hocko Acked-by: Johannes Weiner Cc: Christoph Hellwig Cc: Mel Gorman Cc: Vlastimil Babka Signed-off-by: Andrew Morton --- diff --git a/drivers/tty/sysrq.c b/drivers/tty/sysrq.c index d2b2720db6ca..c6b2b42d2367 100644 --- a/drivers/tty/sysrq.c +++ b/drivers/tty/sysrq.c @@ -342,7 +342,7 @@ static const struct sysrq_key_op sysrq_ftrace_dump_op = { static void sysrq_handle_showmem(int key) { - show_mem(0, NULL); + show_mem(0, NULL, GFP_HIGHUSER_MOVABLE); } static const struct sysrq_key_op sysrq_showmem_op = { .handler = sysrq_handle_showmem, diff --git a/drivers/tty/vt/keyboard.c b/drivers/tty/vt/keyboard.c index be8313cdbac3..9ceeea531cf6 100644 --- a/drivers/tty/vt/keyboard.c +++ b/drivers/tty/vt/keyboard.c @@ -606,7 +606,7 @@ static void fn_scroll_back(struct vc_data *vc) static void fn_show_mem(struct vc_data *vc) { - show_mem(0, NULL); + show_mem(0, NULL, GFP_HIGHUSER_MOVABLE); } static void fn_show_state(struct vc_data *vc) diff --git a/include/linux/mm.h b/include/linux/mm.h index 55cf92c0eb4f..7b316c95cd70 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -1842,7 +1842,7 @@ extern void pagefault_out_of_memory(void); */ #define SHOW_MEM_FILTER_NODES (0x0001u) /* disallowed nodes */ -extern void show_free_areas(unsigned int flags, nodemask_t *nodemask); +extern void show_free_areas(unsigned int flags, nodemask_t *nodemask, gfp_t gfp_mask); #ifdef CONFIG_MMU extern bool can_do_mlock(void); @@ -2582,7 +2582,7 @@ extern void calculate_min_free_kbytes(void); extern int __meminit init_per_zone_wmark_min(void); extern void mem_init(void); extern void __init mmap_init(void); -extern void show_mem(unsigned int flags, nodemask_t *nodemask); +extern void show_mem(unsigned int flags, nodemask_t *nodemask, gfp_t gfp_mask); extern long si_mem_available(void); extern void si_meminfo(struct sysinfo * val); extern void si_meminfo_node(struct sysinfo *val, int nid); diff --git a/init/initramfs.c b/init/initramfs.c index 18229cfe8906..ec966cb78363 100644 --- a/init/initramfs.c +++ b/init/initramfs.c @@ -63,7 +63,7 @@ static void panic_show_mem(const char *fmt, ...) { va_list args; - show_mem(0, NULL); + show_mem(0, NULL, GFP_HIGHUSER_MOVABLE); va_start(args, fmt); panic(fmt, args); va_end(args); diff --git a/kernel/panic.c b/kernel/panic.c index c6eb8f8db0c0..096a1c850277 100644 --- a/kernel/panic.c +++ b/kernel/panic.c @@ -187,7 +187,7 @@ static void panic_print_sys_info(bool console_flush) show_state(); if (panic_print & PANIC_PRINT_MEM_INFO) - show_mem(0, NULL); + show_mem(0, NULL, GFP_HIGHUSER_MOVABLE); if (panic_print & PANIC_PRINT_TIMER_INFO) sysrq_timer_list_show(); diff --git a/lib/show_mem.c b/lib/show_mem.c index 1c26c14ffbb9..b97461a6c4bb 100644 --- a/lib/show_mem.c +++ b/lib/show_mem.c @@ -8,13 +8,13 @@ #include #include -void show_mem(unsigned int filter, nodemask_t *nodemask) +void show_mem(unsigned int filter, nodemask_t *nodemask, gfp_t gfp_mask) { pg_data_t *pgdat; unsigned long total = 0, reserved = 0, highmem = 0; printk("Mem-Info:\n"); - show_free_areas(filter, nodemask); + show_free_areas(filter, nodemask, gfp_mask); for_each_online_pgdat(pgdat) { int zoneid; diff --git a/mm/nommu.c b/mm/nommu.c index 214c70e1d059..8d9c51b6ed8c 100644 --- a/mm/nommu.c +++ b/mm/nommu.c @@ -1030,7 +1030,7 @@ error_free: enomem: pr_err("Allocation of length %lu from process %d (%s) failed\n", len, current->pid, current->comm); - show_free_areas(0, NULL); + show_free_areas(0, NULL, GFP_KERNEL); return -ENOMEM; } @@ -1259,13 +1259,13 @@ error_getting_vma: kmem_cache_free(vm_region_jar, region); pr_warn("Allocation of vma for %lu byte allocation from process %d failed\n", len, current->pid); - show_free_areas(0, NULL); + show_free_areas(0, NULL, GFP_KERNEL); return -ENOMEM; error_getting_region: pr_warn("Allocation of vm region for %lu byte allocation from process %d failed\n", len, current->pid); - show_free_areas(0, NULL); + show_free_areas(0, NULL, GFP_KERNEL); return -ENOMEM; error_maple_preallocate: diff --git a/mm/oom_kill.c b/mm/oom_kill.c index 62975225db69..3e83b6d81af4 100644 --- a/mm/oom_kill.c +++ b/mm/oom_kill.c @@ -461,7 +461,7 @@ static void dump_header(struct oom_control *oc, struct task_struct *p) if (is_memcg_oom(oc)) mem_cgroup_print_oom_meminfo(oc->memcg); else { - show_mem(SHOW_MEM_FILTER_NODES, oc->nodemask); + show_mem(SHOW_MEM_FILTER_NODES, oc->nodemask, oc->gfp_mask); if (should_dump_unreclaim_slab()) dump_unreclaimable_slab(); } diff --git a/mm/page_alloc.c b/mm/page_alloc.c index db434a855a40..405da48c63d8 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -4329,7 +4329,7 @@ static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask) if (!in_task() || !(gfp_mask & __GFP_DIRECT_RECLAIM)) filter &= ~SHOW_MEM_FILTER_NODES; - show_mem(filter, nodemask); + show_mem(filter, nodemask, gfp_mask); } void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...) @@ -6049,6 +6049,15 @@ static void show_migration_types(unsigned char type) printk(KERN_CONT "(%s) ", tmp); } +static bool node_has_managed_zones(pg_data_t *pgdat, int max_zone_idx) +{ + int zone_idx; + for (zone_idx = 0; zone_idx <= max_zone_idx; zone_idx++) + if (zone_managed_pages(pgdat->node_zones + zone_idx)) + return true; + return false; +} + /* * Show free area list (used inside shift_scroll-lock stuff) * We also calculate the percentage fragmentation. We do this by counting the @@ -6058,14 +6067,17 @@ static void show_migration_types(unsigned char type) * SHOW_MEM_FILTER_NODES: suppress nodes that are not allowed by current's * cpuset. */ -void show_free_areas(unsigned int filter, nodemask_t *nodemask) +void show_free_areas(unsigned int filter, nodemask_t *nodemask, gfp_t gfp_mask) { unsigned long free_pcp = 0; + int max_zone_idx = gfp_zone(gfp_mask); int cpu, nid; struct zone *zone; pg_data_t *pgdat; for_each_populated_zone(zone) { + if (zone_idx(zone) > max_zone_idx) + continue; if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) continue; @@ -6103,6 +6115,8 @@ void show_free_areas(unsigned int filter, nodemask_t *nodemask) for_each_online_pgdat(pgdat) { if (show_mem_node_skip(filter, pgdat->node_id, nodemask)) continue; + if (!node_has_managed_zones(pgdat, max_zone_idx)) + continue; printk("Node %d" " active_anon:%lukB" @@ -6159,6 +6173,8 @@ void show_free_areas(unsigned int filter, nodemask_t *nodemask) for_each_populated_zone(zone) { int i; + if (zone_idx(zone) > max_zone_idx) + continue; if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) continue; @@ -6220,6 +6236,8 @@ void show_free_areas(unsigned int filter, nodemask_t *nodemask) unsigned long nr[MAX_ORDER], flags, total = 0; unsigned char types[MAX_ORDER]; + if (zone_idx(zone) > max_zone_idx) + continue; if (show_mem_node_skip(filter, zone_to_nid(zone), nodemask)) continue; show_node(zone);