This patch reduces repetition of set_task_reclaim_state() around
do_try_to_free_pages().

Signed-off-by: Jacob Wen <jian.w....@oracle.com>
---
 mm/vmscan.c | 27 ++++++++++++++++-----------
 1 file changed, 16 insertions(+), 11 deletions(-)

diff --git a/mm/vmscan.c b/mm/vmscan.c
index 257cba79a96d..4bc244b23686 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -3023,6 +3023,10 @@ static unsigned long do_try_to_free_pages(struct 
zonelist *zonelist,
        pg_data_t *last_pgdat;
        struct zoneref *z;
        struct zone *zone;
+       unsigned long ret;
+
+       set_task_reclaim_state(current, &sc->reclaim_state);
+
 retry:
        delayacct_freepages_start();
 
@@ -3069,12 +3073,16 @@ static unsigned long do_try_to_free_pages(struct 
zonelist *zonelist,
 
        delayacct_freepages_end();
 
-       if (sc->nr_reclaimed)
-               return sc->nr_reclaimed;
+       if (sc->nr_reclaimed) {
+               ret = sc->nr_reclaimed;
+               goto out;
+       }
 
        /* Aborted reclaim to try compaction? don't OOM, then */
-       if (sc->compaction_ready)
-               return 1;
+       if (sc->compaction_ready) {
+               ret = 1;
+               goto out;
+       }
 
        /*
         * We make inactive:active ratio decisions based on the node's
@@ -3101,7 +3109,10 @@ static unsigned long do_try_to_free_pages(struct 
zonelist *zonelist,
                goto retry;
        }
 
-       return 0;
+       ret = 0;
+out:
+       set_task_reclaim_state(current, NULL);
+       return ret;
 }
 
 static bool allow_direct_reclaim(pg_data_t *pgdat)
@@ -3269,13 +3280,11 @@ unsigned long try_to_free_pages(struct zonelist 
*zonelist, int order,
        if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask))
                return 1;
 
-       set_task_reclaim_state(current, &sc.reclaim_state);
        trace_mm_vmscan_direct_reclaim_begin(order, sc.gfp_mask);
 
        nr_reclaimed = do_try_to_free_pages(zonelist, &sc);
 
        trace_mm_vmscan_direct_reclaim_end(nr_reclaimed);
-       set_task_reclaim_state(current, NULL);
 
        return nr_reclaimed;
 }
@@ -3347,7 +3356,6 @@ unsigned long try_to_free_mem_cgroup_pages(struct 
mem_cgroup *memcg,
         */
        struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask);
 
-       set_task_reclaim_state(current, &sc.reclaim_state);
        trace_mm_vmscan_memcg_reclaim_begin(0, sc.gfp_mask);
        noreclaim_flag = memalloc_noreclaim_save();
 
@@ -3355,7 +3363,6 @@ unsigned long try_to_free_mem_cgroup_pages(struct 
mem_cgroup *memcg,
 
        memalloc_noreclaim_restore(noreclaim_flag);
        trace_mm_vmscan_memcg_reclaim_end(nr_reclaimed);
-       set_task_reclaim_state(current, NULL);
 
        return nr_reclaimed;
 }
@@ -4023,11 +4030,9 @@ unsigned long shrink_all_memory(unsigned long 
nr_to_reclaim)
 
        fs_reclaim_acquire(sc.gfp_mask);
        noreclaim_flag = memalloc_noreclaim_save();
-       set_task_reclaim_state(current, &sc.reclaim_state);
 
        nr_reclaimed = do_try_to_free_pages(zonelist, &sc);
 
-       set_task_reclaim_state(current, NULL);
        memalloc_noreclaim_restore(noreclaim_flag);
        fs_reclaim_release(sc.gfp_mask);
 
-- 
2.25.1

Reply via email to