]> www.infradead.org Git - users/willy/pagecache.git/commitdiff
mm, swap: simplify percpu cluster updating
authorKairui Song <kasong@tencent.com>
Mon, 13 Jan 2025 17:57:29 +0000 (01:57 +0800)
committerAndrew Morton <akpm@linux-foundation.org>
Sun, 26 Jan 2025 04:22:37 +0000 (20:22 -0800)
Instead of using a returning argument, we can simply store the next
cluster offset to the fixed percpu location, which reduce the stack usage
and simplify the function:

Object size:
./scripts/bloat-o-meter mm/swapfile.o mm/swapfile.o.new
add/remove: 0/0 grow/shrink: 0/2 up/down: 0/-271 (-271)
Function                                     old     new   delta
get_swap_pages                              2847    2733    -114
alloc_swap_scan_cluster                      894     737    -157
Total: Before=30833, After=30562, chg -0.88%

Stack usage:
Before:
swapfile.c:1190:5:get_swap_pages       240    static

After:
swapfile.c:1185:5:get_swap_pages       216    static

Link: https://lkml.kernel.org/r/20250113175732.48099-11-ryncsn@gmail.com
Signed-off-by: Kairui Song <kasong@tencent.com>
Cc: Baoquan He <bhe@redhat.com>
Cc: Barry Song <v-songbaohua@oppo.com>
Cc: Chis Li <chrisl@kernel.org>
Cc: "Huang, Ying" <ying.huang@linux.alibaba.com>
Cc: Hugh Dickens <hughd@google.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Kalesh Singh <kaleshsingh@google.com>
Cc: Nhat Pham <nphamcs@gmail.com>
Cc: Ryan Roberts <ryan.roberts@arm.com>
Cc: Yosry Ahmed <yosryahmed@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
include/linux/swap.h
mm/swapfile.c

index 5fe650beb77d42c80bb8f70b64662f984fe65bed..75b2b0166cb107a34660f66635b2d4f6a606e79a 100644 (file)
@@ -274,9 +274,9 @@ enum swap_cluster_flags {
  * The first page in the swap file is the swap header, which is always marked
  * bad to prevent it from being allocated as an entry. This also prevents the
  * cluster to which it belongs being marked free. Therefore 0 is safe to use as
- * a sentinel to indicate next is not valid in percpu_cluster.
+ * a sentinel to indicate an entry is not valid.
  */
-#define SWAP_NEXT_INVALID      0
+#define SWAP_ENTRY_INVALID     0
 
 #ifdef CONFIG_THP_SWAP
 #define SWAP_NR_ORDERS         (PMD_ORDER + 1)
index 489ac6997a0ceb66377459bb419dc62ac53edd81..6da2f3aa55fbe4e5e482a9eb6d1f4b559b32eb34 100644 (file)
@@ -765,23 +765,23 @@ static bool cluster_alloc_range(struct swap_info_struct *si, struct swap_cluster
        return true;
 }
 
-static unsigned int alloc_swap_scan_cluster(struct swap_info_struct *si, unsigned long offset,
-                                           unsigned int *foundp, unsigned int order,
+/* Try use a new cluster for current CPU and allocate from it. */
+static unsigned int alloc_swap_scan_cluster(struct swap_info_struct *si,
+                                           struct swap_cluster_info *ci,
+                                           unsigned long offset,
+                                           unsigned int order,
                                            unsigned char usage)
 {
-       unsigned long start = offset & ~(SWAPFILE_CLUSTER - 1);
+       unsigned int next = SWAP_ENTRY_INVALID, found = SWAP_ENTRY_INVALID;
+       unsigned long start = ALIGN_DOWN(offset, SWAPFILE_CLUSTER);
        unsigned long end = min(start + SWAPFILE_CLUSTER, si->max);
        unsigned int nr_pages = 1 << order;
        bool need_reclaim, ret;
-       struct swap_cluster_info *ci;
 
-       ci = &si->cluster_info[offset / SWAPFILE_CLUSTER];
        lockdep_assert_held(&ci->lock);
 
-       if (end < nr_pages || ci->count + nr_pages > SWAPFILE_CLUSTER) {
-               offset = SWAP_NEXT_INVALID;
+       if (end < nr_pages || ci->count + nr_pages > SWAPFILE_CLUSTER)
                goto out;
-       }
 
        for (end -= nr_pages; offset <= end; offset += nr_pages) {
                need_reclaim = false;
@@ -795,34 +795,27 @@ static unsigned int alloc_swap_scan_cluster(struct swap_info_struct *si, unsigne
                         * cluster has no flag set, and change of list
                         * won't cause fragmentation.
                         */
-                       if (!cluster_is_usable(ci, order)) {
-                               offset = SWAP_NEXT_INVALID;
+                       if (!cluster_is_usable(ci, order))
                                goto out;
-                       }
                        if (cluster_is_empty(ci))
                                offset = start;
                        /* Reclaim failed but cluster is usable, try next */
                        if (!ret)
                                continue;
                }
-               if (!cluster_alloc_range(si, ci, offset, usage, order)) {
-                       offset = SWAP_NEXT_INVALID;
-                       goto out;
-               }
-               *foundp = offset;
-               if (ci->count == SWAPFILE_CLUSTER) {
-                       offset = SWAP_NEXT_INVALID;
-                       goto out;
-               }
+               if (!cluster_alloc_range(si, ci, offset, usage, order))
+                       break;
+               found = offset;
                offset += nr_pages;
+               if (ci->count < SWAPFILE_CLUSTER && offset <= end)
+                       next = offset;
                break;
        }
-       if (offset > end)
-               offset = SWAP_NEXT_INVALID;
 out:
        relocate_cluster(si, ci);
        unlock_cluster(ci);
-       return offset;
+       __this_cpu_write(si->percpu_cluster->next[order], next);
+       return found;
 }
 
 /* Return true if reclaimed a whole cluster */
@@ -891,8 +884,8 @@ static unsigned long cluster_alloc_swap_entry(struct swap_info_struct *si, int o
                if (cluster_is_usable(ci, order)) {
                        if (cluster_is_empty(ci))
                                offset = cluster_offset(si, ci);
-                       offset = alloc_swap_scan_cluster(si, offset, &found,
-                                                        order, usage);
+                       found = alloc_swap_scan_cluster(si, ci, offset,
+                                                       order, usage);
                } else {
                        unlock_cluster(ci);
                }
@@ -903,8 +896,8 @@ static unsigned long cluster_alloc_swap_entry(struct swap_info_struct *si, int o
 new_cluster:
        ci = isolate_lock_cluster(si, &si->free_clusters);
        if (ci) {
-               offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci),
-                                                &found, order, usage);
+               found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci),
+                                               order, usage);
                if (found)
                        goto done;
        }
@@ -917,8 +910,8 @@ new_cluster:
                unsigned int frags = 0, frags_existing;
 
                while ((ci = isolate_lock_cluster(si, &si->nonfull_clusters[order]))) {
-                       offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci),
-                                                        &found, order, usage);
+                       found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci),
+                                                       order, usage);
                        if (found)
                                goto done;
                        /* Clusters failed to allocate are moved to frag_clusters */
@@ -935,8 +928,8 @@ new_cluster:
                         * per-CPU usage, but they could contain newly released
                         * reclaimable (eg. lazy-freed swap cache) slots.
                         */
-                       offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci),
-                                                        &found, order, usage);
+                       found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci),
+                                                       order, usage);
                        if (found)
                                goto done;
                        frags++;
@@ -962,21 +955,20 @@ new_cluster:
                 */
                while ((ci = isolate_lock_cluster(si, &si->frag_clusters[o]))) {
                        atomic_long_dec(&si->frag_cluster_nr[o]);
-                       offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci),
-                                                        &found, order, usage);
+                       found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci),
+                                                       0, usage);
                        if (found)
                                goto done;
                }
 
                while ((ci = isolate_lock_cluster(si, &si->nonfull_clusters[o]))) {
-                       offset = alloc_swap_scan_cluster(si, cluster_offset(si, ci),
-                                                        &found, order, usage);
+                       found = alloc_swap_scan_cluster(si, ci, cluster_offset(si, ci),
+                                                       0, usage);
                        if (found)
                                goto done;
                }
        }
 done:
-       __this_cpu_write(si->percpu_cluster->next[order], offset);
        local_unlock(&si->percpu_cluster->lock);
 
        return found;
@@ -3200,7 +3192,7 @@ static struct swap_cluster_info *setup_clusters(struct swap_info_struct *si,
 
                cluster = per_cpu_ptr(si->percpu_cluster, cpu);
                for (i = 0; i < SWAP_NR_ORDERS; i++)
-                       cluster->next[i] = SWAP_NEXT_INVALID;
+                       cluster->next[i] = SWAP_ENTRY_INVALID;
                local_lock_init(&cluster->lock);
        }