On Thu 27-06-19 20:54:03, Minchan Kim wrote:
> The isolate counting is pecpu counter so it would be not huge gain
> to work them by batch. Rather than complicating to make them batch,
> let's make it more stright-foward via adding the counting logic
> into [isolate|putback]_lru_page API.
> 
> * v1
>  * fix accounting bug - Hillf
> 
> Link: http://lkml.kernel.org/r/20190531165927.ga20...@cmpxchg.org
> Suggested-by: Johannes Weiner <han...@cmpxchg.org>
> Signed-off-by: Minchan Kim <minc...@kernel.org>

I like that the NR_ISOLATED_$FOO handling gets out of any code except
for vmscan and migration. This is definitely an improvement.

I haven't spotted any imbalance so I hope I haven't really missed any
path.

Acked-by: Michal Hocko <mho...@suse.com>

Thanks!

> ---
>  mm/compaction.c     |  2 --
>  mm/gup.c            |  7 +------
>  mm/khugepaged.c     |  3 ---
>  mm/memory-failure.c |  3 ---
>  mm/memory_hotplug.c |  4 ----
>  mm/mempolicy.c      |  6 +-----
>  mm/migrate.c        | 37 ++++++++-----------------------------
>  mm/vmscan.c         | 22 ++++++++++++++++------
>  8 files changed, 26 insertions(+), 58 deletions(-)
> 
> diff --git a/mm/compaction.c b/mm/compaction.c
> index 9e1b9acb116b..c6591682deda 100644
> --- a/mm/compaction.c
> +++ b/mm/compaction.c
> @@ -982,8 +982,6 @@ isolate_migratepages_block(struct compact_control *cc, 
> unsigned long low_pfn,
>  
>               /* Successfully isolated */
>               del_page_from_lru_list(page, lruvec, page_lru(page));
> -             inc_node_page_state(page,
> -                             NR_ISOLATED_ANON + page_is_file_cache(page));
>  
>  isolate_success:
>               list_add(&page->lru, &cc->migratepages);
> diff --git a/mm/gup.c b/mm/gup.c
> index 7dde2e3a1963..aec3a2b7e61b 100644
> --- a/mm/gup.c
> +++ b/mm/gup.c
> @@ -1473,13 +1473,8 @@ static long check_and_migrate_cma_pages(struct 
> task_struct *tsk,
>                                       drain_allow = false;
>                               }
>  
> -                             if (!isolate_lru_page(head)) {
> +                             if (!isolate_lru_page(head))
>                                       list_add_tail(&head->lru, 
> &cma_page_list);
> -                                     mod_node_page_state(page_pgdat(head),
> -                                                         NR_ISOLATED_ANON +
> -                                                         
> page_is_file_cache(head),
> -                                                         
> hpage_nr_pages(head));
> -                             }
>                       }
>               }
>       }
> diff --git a/mm/khugepaged.c b/mm/khugepaged.c
> index 0f7419938008..7da34e198ec5 100644
> --- a/mm/khugepaged.c
> +++ b/mm/khugepaged.c
> @@ -503,7 +503,6 @@ void __khugepaged_exit(struct mm_struct *mm)
>  
>  static void release_pte_page(struct page *page)
>  {
> -     dec_node_page_state(page, NR_ISOLATED_ANON + page_is_file_cache(page));
>       unlock_page(page);
>       putback_lru_page(page);
>  }
> @@ -602,8 +601,6 @@ static int __collapse_huge_page_isolate(struct 
> vm_area_struct *vma,
>                       result = SCAN_DEL_PAGE_LRU;
>                       goto out;
>               }
> -             inc_node_page_state(page,
> -                             NR_ISOLATED_ANON + page_is_file_cache(page));
>               VM_BUG_ON_PAGE(!PageLocked(page), page);
>               VM_BUG_ON_PAGE(PageLRU(page), page);
>  
> diff --git a/mm/memory-failure.c b/mm/memory-failure.c
> index 7e08cbf3ba49..3586e8226e4e 100644
> --- a/mm/memory-failure.c
> +++ b/mm/memory-failure.c
> @@ -1795,9 +1795,6 @@ static int __soft_offline_page(struct page *page, int 
> flags)
>                * so use !__PageMovable instead for LRU page's mapping
>                * cannot have PAGE_MAPPING_MOVABLE.
>                */
> -             if (!__PageMovable(page))
> -                     inc_node_page_state(page, NR_ISOLATED_ANON +
> -                                             page_is_file_cache(page));
>               list_add(&page->lru, &pagelist);
>               ret = migrate_pages(&pagelist, new_page, NULL, MPOL_MF_MOVE_ALL,
>                                       MIGRATE_SYNC, MR_MEMORY_FAILURE);
> diff --git a/mm/memory_hotplug.c b/mm/memory_hotplug.c
> index dfab21dc33dc..68577c677b46 100644
> --- a/mm/memory_hotplug.c
> +++ b/mm/memory_hotplug.c
> @@ -1384,10 +1384,6 @@ do_migrate_range(unsigned long start_pfn, unsigned 
> long end_pfn)
>                       ret = isolate_movable_page(page, ISOLATE_UNEVICTABLE);
>               if (!ret) { /* Success */
>                       list_add_tail(&page->lru, &source);
> -                     if (!__PageMovable(page))
> -                             inc_node_page_state(page, NR_ISOLATED_ANON +
> -                                                 page_is_file_cache(page));
> -
>               } else {
>                       pr_warn("failed to isolate pfn %lx\n", pfn);
>                       dump_page(page, "isolation failed");
> diff --git a/mm/mempolicy.c b/mm/mempolicy.c
> index 64562809bf3b..03081f3404ca 100644
> --- a/mm/mempolicy.c
> +++ b/mm/mempolicy.c
> @@ -994,12 +994,8 @@ static int migrate_page_add(struct page *page, struct 
> list_head *pagelist,
>        * Avoid migrating a page that is shared with others.
>        */
>       if ((flags & MPOL_MF_MOVE_ALL) || page_mapcount(head) == 1) {
> -             if (!isolate_lru_page(head)) {
> +             if (!isolate_lru_page(head))
>                       list_add_tail(&head->lru, pagelist);
> -                     mod_node_page_state(page_pgdat(head),
> -                             NR_ISOLATED_ANON + page_is_file_cache(head),
> -                             hpage_nr_pages(head));
> -             }
>       }
>  
>       return 0;
> diff --git a/mm/migrate.c b/mm/migrate.c
> index 572b4bc85d76..5583324c01e7 100644
> --- a/mm/migrate.c
> +++ b/mm/migrate.c
> @@ -190,8 +190,6 @@ void putback_movable_pages(struct list_head *l)
>                       unlock_page(page);
>                       put_page(page);
>               } else {
> -                     mod_node_page_state(page_pgdat(page), NR_ISOLATED_ANON +
> -                                     page_is_file_cache(page), 
> -hpage_nr_pages(page));
>                       putback_lru_page(page);
>               }
>       }
> @@ -1181,10 +1179,17 @@ static ICE_noinline int unmap_and_move(new_page_t 
> get_new_page,
>               return -ENOMEM;
>  
>       if (page_count(page) == 1) {
> +             bool is_lru = !__PageMovable(page);
> +
>               /* page was freed from under us. So we are done. */
>               ClearPageActive(page);
>               ClearPageUnevictable(page);
> -             if (unlikely(__PageMovable(page))) {
> +             if (likely(is_lru))
> +                     mod_node_page_state(page_pgdat(page),
> +                                             NR_ISOLATED_ANON +
> +                                             page_is_file_cache(page),
> +                                             -hpage_nr_pages(page));
> +             else {
>                       lock_page(page);
>                       if (!PageMovable(page))
>                               __ClearPageIsolated(page);
> @@ -1210,15 +1215,6 @@ static ICE_noinline int unmap_and_move(new_page_t 
> get_new_page,
>                * restored.
>                */
>               list_del(&page->lru);
> -
> -             /*
> -              * Compaction can migrate also non-LRU pages which are
> -              * not accounted to NR_ISOLATED_*. They can be recognized
> -              * as __PageMovable
> -              */
> -             if (likely(!__PageMovable(page)))
> -                     mod_node_page_state(page_pgdat(page), NR_ISOLATED_ANON +
> -                                     page_is_file_cache(page), 
> -hpage_nr_pages(page));
>       }
>  
>       /*
> @@ -1572,9 +1568,6 @@ static int add_page_for_migration(struct mm_struct *mm, 
> unsigned long addr,
>  
>               err = 0;
>               list_add_tail(&head->lru, pagelist);
> -             mod_node_page_state(page_pgdat(head),
> -                     NR_ISOLATED_ANON + page_is_file_cache(head),
> -                     hpage_nr_pages(head));
>       }
>  out_putpage:
>       /*
> @@ -1890,8 +1883,6 @@ static struct page *alloc_misplaced_dst_page(struct 
> page *page,
>  
>  static int numamigrate_isolate_page(pg_data_t *pgdat, struct page *page)
>  {
> -     int page_lru;
> -
>       VM_BUG_ON_PAGE(compound_order(page) && !PageTransHuge(page), page);
>  
>       /* Avoid migrating to a node that is nearly full */
> @@ -1913,10 +1904,6 @@ static int numamigrate_isolate_page(pg_data_t *pgdat, 
> struct page *page)
>               return 0;
>       }
>  
> -     page_lru = page_is_file_cache(page);
> -     mod_node_page_state(page_pgdat(page), NR_ISOLATED_ANON + page_lru,
> -                             hpage_nr_pages(page));
> -
>       /*
>        * Isolating the page has taken another reference, so the
>        * caller's reference can be safely dropped without the page
> @@ -1971,8 +1958,6 @@ int migrate_misplaced_page(struct page *page, struct 
> vm_area_struct *vma,
>       if (nr_remaining) {
>               if (!list_empty(&migratepages)) {
>                       list_del(&page->lru);
> -                     dec_node_page_state(page, NR_ISOLATED_ANON +
> -                                     page_is_file_cache(page));
>                       putback_lru_page(page);
>               }
>               isolated = 0;
> @@ -2002,7 +1987,6 @@ int migrate_misplaced_transhuge_page(struct mm_struct 
> *mm,
>       pg_data_t *pgdat = NODE_DATA(node);
>       int isolated = 0;
>       struct page *new_page = NULL;
> -     int page_lru = page_is_file_cache(page);
>       unsigned long start = address & HPAGE_PMD_MASK;
>  
>       new_page = alloc_pages_node(node,
> @@ -2048,8 +2032,6 @@ int migrate_misplaced_transhuge_page(struct mm_struct 
> *mm,
>               /* Retake the callers reference and putback on LRU */
>               get_page(page);
>               putback_lru_page(page);
> -             mod_node_page_state(page_pgdat(page),
> -                      NR_ISOLATED_ANON + page_lru, -HPAGE_PMD_NR);
>  
>               goto out_unlock;
>       }
> @@ -2099,9 +2081,6 @@ int migrate_misplaced_transhuge_page(struct mm_struct 
> *mm,
>       count_vm_events(PGMIGRATE_SUCCESS, HPAGE_PMD_NR);
>       count_vm_numa_events(NUMA_PAGE_MIGRATE, HPAGE_PMD_NR);
>  
> -     mod_node_page_state(page_pgdat(page),
> -                     NR_ISOLATED_ANON + page_lru,
> -                     -HPAGE_PMD_NR);
>       return isolated;
>  
>  out_fail:
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index 49e9ee4d771d..223ce5da08f0 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -1014,6 +1014,9 @@ int remove_mapping(struct address_space *mapping, 
> struct page *page)
>  void putback_lru_page(struct page *page)
>  {
>       lru_cache_add(page);
> +     mod_node_page_state(page_pgdat(page),
> +                             NR_ISOLATED_ANON + page_is_file_cache(page),
> +                             -hpage_nr_pages(page));
>       put_page(page);         /* drop ref from isolate */
>  }
>  
> @@ -1479,6 +1482,9 @@ static unsigned long shrink_page_list(struct list_head 
> *page_list,
>                */
>               nr_reclaimed += nr_pages;
>  
> +             mod_node_page_state(pgdat, NR_ISOLATED_ANON +
> +                                             page_is_file_cache(page),
> +                                             -nr_pages);
>               /*
>                * Is there need to periodically free_page_list? It would
>                * appear not as the counts should be low
> @@ -1554,7 +1560,6 @@ unsigned long reclaim_clean_pages_from_list(struct zone 
> *zone,
>       ret = shrink_page_list(&clean_pages, zone->zone_pgdat, &sc,
>                       TTU_IGNORE_ACCESS, &dummy_stat, true);
>       list_splice(&clean_pages, page_list);
> -     mod_node_page_state(zone->zone_pgdat, NR_ISOLATED_FILE, -ret);
>       return ret;
>  }
>  
> @@ -1630,6 +1635,9 @@ int __isolate_lru_page(struct page *page, 
> isolate_mode_t mode)
>                */
>               ClearPageLRU(page);
>               ret = 0;
> +             __mod_node_page_state(page_pgdat(page), NR_ISOLATED_ANON +
> +                                             page_is_file_cache(page),
> +                                             hpage_nr_pages(page));
>       }
>  
>       return ret;
> @@ -1761,6 +1769,7 @@ static unsigned long isolate_lru_pages(unsigned long 
> nr_to_scan,
>       trace_mm_vmscan_lru_isolate(sc->reclaim_idx, sc->order, nr_to_scan,
>                                   total_scan, skipped, nr_taken, mode, lru);
>       update_lru_sizes(lruvec, lru, nr_zone_taken);
> +
>       return nr_taken;
>  }
>  
> @@ -1809,6 +1818,9 @@ int isolate_lru_page(struct page *page)
>                       ClearPageLRU(page);
>                       del_page_from_lru_list(page, lruvec, lru);
>                       ret = 0;
> +                     mod_node_page_state(pgdat, NR_ISOLATED_ANON +
> +                                             page_is_file_cache(page),
> +                                             hpage_nr_pages(page));
>               }
>               spin_unlock_irq(&pgdat->lru_lock);
>       }
> @@ -1900,6 +1912,9 @@ static unsigned noinline_for_stack 
> move_pages_to_lru(struct lruvec *lruvec,
>               update_lru_size(lruvec, lru, page_zonenum(page), nr_pages);
>               list_move(&page->lru, &lruvec->lists[lru]);
>  
> +             __mod_node_page_state(pgdat, NR_ISOLATED_ANON +
> +                                             page_is_file_cache(page),
> +                                             -hpage_nr_pages(page));
>               if (put_page_testzero(page)) {
>                       __ClearPageLRU(page);
>                       __ClearPageActive(page);
> @@ -1977,7 +1992,6 @@ shrink_inactive_list(unsigned long nr_to_scan, struct 
> lruvec *lruvec,
>       nr_taken = isolate_lru_pages(nr_to_scan, lruvec, &page_list,
>                                    &nr_scanned, sc, lru);
>  
> -     __mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, nr_taken);
>       reclaim_stat->recent_scanned[file] += nr_taken;
>  
>       item = current_is_kswapd() ? PGSCAN_KSWAPD : PGSCAN_DIRECT;
> @@ -2003,8 +2017,6 @@ shrink_inactive_list(unsigned long nr_to_scan, struct 
> lruvec *lruvec,
>  
>       move_pages_to_lru(lruvec, &page_list);
>  
> -     __mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, -nr_taken);
> -
>       spin_unlock_irq(&pgdat->lru_lock);
>  
>       mem_cgroup_uncharge_list(&page_list);
> @@ -2063,7 +2075,6 @@ static void shrink_active_list(unsigned long nr_to_scan,
>       nr_taken = isolate_lru_pages(nr_to_scan, lruvec, &l_hold,
>                                    &nr_scanned, sc, lru);
>  
> -     __mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, nr_taken);
>       reclaim_stat->recent_scanned[file] += nr_taken;
>  
>       __count_vm_events(PGREFILL, nr_scanned);
> @@ -2132,7 +2143,6 @@ static void shrink_active_list(unsigned long nr_to_scan,
>       __count_vm_events(PGDEACTIVATE, nr_deactivate);
>       __count_memcg_events(lruvec_memcg(lruvec), PGDEACTIVATE, nr_deactivate);
>  
> -     __mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, -nr_taken);
>       spin_unlock_irq(&pgdat->lru_lock);
>  
>       mem_cgroup_uncharge_list(&l_active);
> -- 
> 2.22.0.410.gd8fdbe21b5-goog

-- 
Michal Hocko
SUSE Labs

Reply via email to