On 11/14/2014 08:05 AM, Joonsoo Kim wrote: >> What about this scenario, with pageblock order: >> >> - record cc->migrate_pfn pointing to pageblock X >> - isolate_migratepages() skips the pageblock due to e.g. skip bit, >> or the pageblock being a THP already... >> - loop to pageblock X+1, last_migrated_pfn is still set to pfn of >> pageblock X (more precisely the pfn is (X << pageblock_order) - 1 >> per your code, but doesn't matter) >> - isolate_migratepages isolates something, but ends up somewhere in >> the middle of pageblock due to COMPACT_CLUSTER_MAX >> - cc->migrate_pfn points to pageblock X+1 (plus some pages it scanned) >> - so it will decide that it has fully migrated pageblock X and it's >> time to drain. But the drain is most likely useless - we didn't >> migrate anything in pageblock X, we skipped it. And in X+1 we didn't >> migrate everything yet, so we should drain only after finishing the >> other part of the pageblock. > > Yes, but, it can be easily fixed. > > while (compact_finished()) { > unsigned long prev_migrate_pfn = cc->migrate_pfn; > > isolate_migratepages() > switch case { > NONE: > goto check_drain; > SUCCESS: > if (!last_migrated_pfn) > last_migrated_pfn = prev_migrate_pfn; > } > > ... > > check_drain: (at the end of loop) > ... > }
Good suggestion, also gets rid of the awkward subtraction of 1 in the current patch. Thanks. >> In short, "last_migrated_pfn" is not "last position of migrate >> scanner" but "last block where we *actually* migrated". > > Okay. Now I get it. > Nevertheless, I'd like to change logic like above. > > One problem of your approach is that it can't detect some cases. > > Let's think about following case. > '|' denotes aligned block boundary. > '^' denotes migrate_pfn at certain time. > > Assume that last_migrated_pfn = 0; > > |--------------|-------------|--------------| > ^ ^ > before isolate after isolate > > In this case, your code just records position of second '^' to > last_migrated_pfn and skip to flush. But, flush is needed if we > migrate some pages because we move away from previous aligned block. > > Thanks. > Right, so the patch below implements your suggestion, and the last_migrated_pfn initialization fix. I named the variable "isolate_start_pfn" instead of prev_migrate_pfn, as it's where the migrate scanner isolation starts, and having both prev_migrate_pfn and last_migrated_pfn would be more confusing I think. ------8<------ From: Vlastimil Babka <vba...@suse.cz> Date: Mon, 3 Nov 2014 15:28:01 +0100 Subject: [PATCH] mm-compaction-more-focused-lru-and-pcplists-draining-fix As Joonsoo Kim pointed out, last_migrate_pfn was reset to 0 by mistake at each iteration in compact_zone(). This mistake could result in fail to recognize immediately draining points for orders smaller than pageblock. Joonsoo has also suggested an improvement to detecting cc->order aligned block where migration might have occured - before this fix, some of the drain opportunities might have been missed. Signed-off-by: Vlastimil Babka <vba...@suse.cz> Cc: Minchan Kim <minc...@kernel.org> Cc: Mel Gorman <mgor...@suse.de> Cc: Joonsoo Kim <iamjoonsoo....@lge.com> Cc: Michal Nazarewicz <min...@mina86.com> Cc: Naoya Horiguchi <n-horigu...@ah.jp.nec.com> Cc: Christoph Lameter <c...@linux.com> Cc: Rik van Riel <r...@redhat.com> Cc: David Rientjes <rient...@google.com> --- mm/compaction.c | 24 +++++++++++++----------- 1 file changed, 13 insertions(+), 11 deletions(-) diff --git a/mm/compaction.c b/mm/compaction.c index fe43e60..100e6e8 100644 --- a/mm/compaction.c +++ b/mm/compaction.c @@ -1144,6 +1144,7 @@ static int compact_zone(struct zone *zone, struct compact_control *cc) unsigned long end_pfn = zone_end_pfn(zone); const int migratetype = gfpflags_to_migratetype(cc->gfp_mask); const bool sync = cc->mode != MIGRATE_ASYNC; + unsigned long last_migrated_pfn = 0; ret = compaction_suitable(zone, cc->order, cc->alloc_flags, cc->classzone_idx); @@ -1189,7 +1190,7 @@ static int compact_zone(struct zone *zone, struct compact_control *cc) while ((ret = compact_finished(zone, cc, migratetype)) == COMPACT_CONTINUE) { int err; - unsigned long last_migrated_pfn = 0; + unsigned long isolate_start_pfn = cc->migrate_pfn; switch (isolate_migratepages(zone, cc)) { case ISOLATE_ABORT: @@ -1230,21 +1231,22 @@ static int compact_zone(struct zone *zone, struct compact_control *cc) } /* - * Record where we have freed pages by migration and not yet - * flushed them to buddy allocator. Subtract 1, because often - * we finish a pageblock and migrate_pfn points to the first - * page* of the next one. In that case we want the drain below - * to happen immediately. + * Record where we could have freed pages by migration and not + * yet flushed them to buddy allocator. We use the pfn that + * isolate_migratepages() started from in this loop iteration + * - this is the lowest page that could have been isolated and + * then freed by migration. */ if (!last_migrated_pfn) - last_migrated_pfn = cc->migrate_pfn - 1; + last_migrated_pfn = isolate_start_pfn; check_drain: /* - * Have we moved away from the previous cc->order aligned block - * where we migrated from? If yes, flush the pages that were - * freed, so that they can merge and compact_finished() can - * detect immediately if allocation should succeed. + * Has the migration scanner moved away from the previous + * cc->order aligned block where we migrated from? If yes, + * flush the pages that were freed, so that they can merge and + * compact_finished() can detect immediately if allocation + * would succeed. */ if (cc->order > 0 && last_migrated_pfn) { int cpu; -- 2.1.2 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/