]> www.infradead.org Git - users/jedix/linux-maple.git/commitdiff
mm/page-writeback: consolidate wb_thresh bumping logic into __wb_calc_thresh
authorJim Zhao <jimzhao.ai@gmail.com>
Thu, 21 Nov 2024 10:05:39 +0000 (18:05 +0800)
committerAndrew Morton <akpm@linux-foundation.org>
Sun, 26 Jan 2025 04:22:45 +0000 (20:22 -0800)
Address the feedback from 39ac99852fca ("mm/page-writeback: raise
wb_thresh to prevent write blocking with strictlimit)".  The wb_thresh
bumping logic is scattered across wb_position_ratio, __wb_calc_thresh, and
wb_update_dirty_ratelimit.  For consistency, consolidate all wb_thresh
bumping logic into __wb_calc_thresh.

Link: https://lkml.kernel.org/r/20241121100539.605818-1-jimzhao.ai@gmail.com
Signed-off-by: Jim Zhao <jimzhao.ai@gmail.com>
Reviewed-by: Jan Kara <jack@suse.cz>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Kemeng Shi <shikemeng@huaweicloud.com>
Cc: Guenter Roeck <linux@roeck-us.net>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
mm/page-writeback.c

index d9861e42b2bd2431e3a5538061bfe49821c13826..4f5970723cf2a775bb98f63e1f0e83b1c9cc7ce4 100644 (file)
@@ -942,26 +942,25 @@ static unsigned long __wb_calc_thresh(struct dirty_throttle_control *dtc,
        wb_min_max_ratio(wb, &wb_min_ratio, &wb_max_ratio);
 
        wb_thresh += (thresh * wb_min_ratio) / (100 * BDI_RATIO_SCALE);
-       wb_max_thresh = thresh * wb_max_ratio / (100 * BDI_RATIO_SCALE);
-       if (wb_thresh > wb_max_thresh)
-               wb_thresh = wb_max_thresh;
 
        /*
-        * With strictlimit flag, the wb_thresh is treated as
-        * a hard limit in balance_dirty_pages() and wb_position_ratio().
-        * It's possible that wb_thresh is close to zero, not because
-        * the device is slow, but because it has been inactive.
-        * To prevent occasional writes from being blocked, we raise wb_thresh.
+        * It's very possible that wb_thresh is close to 0 not because the
+        * device is slow, but that it has remained inactive for long time.
+        * Honour such devices a reasonable good (hopefully IO efficient)
+        * threshold, so that the occasional writes won't be blocked and active
+        * writes can rampup the threshold quickly.
         */
-       if (unlikely(wb->bdi->capabilities & BDI_CAP_STRICTLIMIT)) {
-               unsigned long limit = hard_dirty_limit(dom, dtc->thresh);
-               u64 wb_scale_thresh = 0;
-
-               if (limit > dtc->dirty)
-                       wb_scale_thresh = (limit - dtc->dirty) / 100;
-               wb_thresh = max(wb_thresh, min(wb_scale_thresh, wb_max_thresh / 4));
+       if (thresh > dtc->dirty) {
+               if (unlikely(wb->bdi->capabilities & BDI_CAP_STRICTLIMIT))
+                       wb_thresh = max(wb_thresh, (thresh - dtc->dirty) / 100);
+               else
+                       wb_thresh = max(wb_thresh, (thresh - dtc->dirty) / 8);
        }
 
+       wb_max_thresh = thresh * wb_max_ratio / (100 * BDI_RATIO_SCALE);
+       if (wb_thresh > wb_max_thresh)
+               wb_thresh = wb_max_thresh;
+
        return wb_thresh;
 }
 
@@ -969,6 +968,7 @@ unsigned long wb_calc_thresh(struct bdi_writeback *wb, unsigned long thresh)
 {
        struct dirty_throttle_control gdtc = { GDTC_INIT(wb) };
 
+       domain_dirty_avail(&gdtc, true);
        return __wb_calc_thresh(&gdtc, thresh);
 }
 
@@ -1145,12 +1145,6 @@ static void wb_position_ratio(struct dirty_throttle_control *dtc)
        if (unlikely(wb->bdi->capabilities & BDI_CAP_STRICTLIMIT)) {
                long long wb_pos_ratio;
 
-               if (dtc->wb_dirty < 8) {
-                       dtc->pos_ratio = min_t(long long, pos_ratio * 2,
-                                          2 << RATELIMIT_CALC_SHIFT);
-                       return;
-               }
-
                if (dtc->wb_dirty >= wb_thresh)
                        return;
 
@@ -1221,14 +1215,6 @@ static void wb_position_ratio(struct dirty_throttle_control *dtc)
         */
        if (unlikely(wb_thresh > dtc->thresh))
                wb_thresh = dtc->thresh;
-       /*
-        * It's very possible that wb_thresh is close to 0 not because the
-        * device is slow, but that it has remained inactive for long time.
-        * Honour such devices a reasonable good (hopefully IO efficient)
-        * threshold, so that the occasional writes won't be blocked and active
-        * writes can rampup the threshold quickly.
-        */
-       wb_thresh = max(wb_thresh, (limit - dtc->dirty) / 8);
        /*
         * scale global setpoint to wb's:
         *      wb_setpoint = setpoint * wb_thresh / thresh
@@ -1484,17 +1470,10 @@ static void wb_update_dirty_ratelimit(struct dirty_throttle_control *dtc,
         * balanced_dirty_ratelimit = task_ratelimit * write_bw / dirty_rate).
         * Hence, to calculate "step" properly, we have to use wb_dirty as
         * "dirty" and wb_setpoint as "setpoint".
-        *
-        * We rampup dirty_ratelimit forcibly if wb_dirty is low because
-        * it's possible that wb_thresh is close to zero due to inactivity
-        * of backing device.
         */
        if (unlikely(wb->bdi->capabilities & BDI_CAP_STRICTLIMIT)) {
                dirty = dtc->wb_dirty;
-               if (dtc->wb_dirty < 8)
-                       setpoint = dtc->wb_dirty + 1;
-               else
-                       setpoint = (dtc->wb_thresh + dtc->wb_bg_thresh) / 2;
+               setpoint = (dtc->wb_thresh + dtc->wb_bg_thresh) / 2;
        }
 
        if (dirty < setpoint) {