diff options
Diffstat (limited to 'mm/vmscan.c')
| -rw-r--r-- | mm/vmscan.c | 20 | 
1 files changed, 19 insertions, 1 deletions
| diff --git a/mm/vmscan.c b/mm/vmscan.c index 9b6497eda806..32c661d66a45 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -1158,7 +1158,7 @@ unsigned long reclaim_clean_pages_from_list(struct zone *zone,  			TTU_UNMAP|TTU_IGNORE_ACCESS,  			&dummy1, &dummy2, &dummy3, &dummy4, &dummy5, true);  	list_splice(&clean_pages, page_list); -	__mod_zone_page_state(zone, NR_ISOLATED_FILE, -ret); +	mod_zone_page_state(zone, NR_ISOLATED_FILE, -ret);  	return ret;  } @@ -1916,6 +1916,24 @@ static void get_scan_count(struct lruvec *lruvec, struct scan_control *sc,  		get_lru_size(lruvec, LRU_INACTIVE_FILE);  	/* +	 * Prevent the reclaimer from falling into the cache trap: as +	 * cache pages start out inactive, every cache fault will tip +	 * the scan balance towards the file LRU.  And as the file LRU +	 * shrinks, so does the window for rotation from references. +	 * This means we have a runaway feedback loop where a tiny +	 * thrashing file LRU becomes infinitely more attractive than +	 * anon pages.  Try to detect this based on file LRU size. +	 */ +	if (global_reclaim(sc)) { +		unsigned long free = zone_page_state(zone, NR_FREE_PAGES); + +		if (unlikely(file + free <= high_wmark_pages(zone))) { +			scan_balance = SCAN_ANON; +			goto out; +		} +	} + +	/*  	 * There is enough inactive page cache, do not reclaim  	 * anything from the anonymous working set right now.  	 */ | 

