diff options
| -rw-r--r-- | include/linux/memcontrol.h | 6 | ||||
| -rw-r--r-- | mm/filemap.c | 18 | ||||
| -rw-r--r-- | mm/memcontrol.c | 44 | 
3 files changed, 52 insertions, 16 deletions
diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h index f944591765e..3558a5e268c 100644 --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h @@ -122,6 +122,8 @@ struct zone_reclaim_stat*  mem_cgroup_get_reclaim_stat_from_page(struct page *page);  extern void mem_cgroup_print_oom_info(struct mem_cgroup *memcg,  					struct task_struct *p); +extern void mem_cgroup_replace_page_cache(struct page *oldpage, +					struct page *newpage);  #ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP  extern int do_swap_account; @@ -369,6 +371,10 @@ static inline  void mem_cgroup_count_vm_event(struct mm_struct *mm, enum vm_event_item idx)  {  } +static inline void mem_cgroup_replace_page_cache(struct page *oldpage, +				struct page *newpage) +{ +}  #endif /* CONFIG_CGROUP_MEM_CONT */  #if !defined(CONFIG_CGROUP_MEM_RES_CTLR) || !defined(CONFIG_DEBUG_VM) diff --git a/mm/filemap.c b/mm/filemap.c index c4ee2e918be..97f49ed35bd 100644 --- a/mm/filemap.c +++ b/mm/filemap.c @@ -393,24 +393,11 @@ EXPORT_SYMBOL(filemap_write_and_wait_range);  int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask)  {  	int error; -	struct mem_cgroup *memcg = NULL;  	VM_BUG_ON(!PageLocked(old));  	VM_BUG_ON(!PageLocked(new));  	VM_BUG_ON(new->mapping); -	/* -	 * This is not page migration, but prepare_migration and -	 * end_migration does enough work for charge replacement. -	 * -	 * In the longer term we probably want a specialized function -	 * for moving the charge from old to new in a more efficient -	 * manner. -	 */ -	error = mem_cgroup_prepare_migration(old, new, &memcg, gfp_mask); -	if (error) -		return error; -  	error = radix_tree_preload(gfp_mask & ~__GFP_HIGHMEM);  	if (!error) {  		struct address_space *mapping = old->mapping; @@ -432,13 +419,12 @@ int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask)  		if (PageSwapBacked(new))  			__inc_zone_page_state(new, NR_SHMEM);  		spin_unlock_irq(&mapping->tree_lock); +		/* mem_cgroup codes must not be called under tree_lock */ +		mem_cgroup_replace_page_cache(old, new);  		radix_tree_preload_end();  		if (freepage)  			freepage(old);  		page_cache_release(old); -		mem_cgroup_end_migration(memcg, old, new, true); -	} else { -		mem_cgroup_end_migration(memcg, old, new, false);  	}  	return error; diff --git a/mm/memcontrol.c b/mm/memcontrol.c index d87aa3510c5..0b2d4036f1c 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -3432,6 +3432,50 @@ void mem_cgroup_end_migration(struct mem_cgroup *memcg,  	cgroup_release_and_wakeup_rmdir(&memcg->css);  } +/* + * At replace page cache, newpage is not under any memcg but it's on + * LRU. So, this function doesn't touch res_counter but handles LRU + * in correct way. Both pages are locked so we cannot race with uncharge. + */ +void mem_cgroup_replace_page_cache(struct page *oldpage, +				  struct page *newpage) +{ +	struct mem_cgroup *memcg; +	struct page_cgroup *pc; +	struct zone *zone; +	enum charge_type type = MEM_CGROUP_CHARGE_TYPE_CACHE; +	unsigned long flags; + +	if (mem_cgroup_disabled()) +		return; + +	pc = lookup_page_cgroup(oldpage); +	/* fix accounting on old pages */ +	lock_page_cgroup(pc); +	memcg = pc->mem_cgroup; +	mem_cgroup_charge_statistics(memcg, PageCgroupCache(pc), -1); +	ClearPageCgroupUsed(pc); +	unlock_page_cgroup(pc); + +	if (PageSwapBacked(oldpage)) +		type = MEM_CGROUP_CHARGE_TYPE_SHMEM; + +	zone = page_zone(newpage); +	pc = lookup_page_cgroup(newpage); +	/* +	 * Even if newpage->mapping was NULL before starting replacement, +	 * the newpage may be on LRU(or pagevec for LRU) already. We lock +	 * LRU while we overwrite pc->mem_cgroup. +	 */ +	spin_lock_irqsave(&zone->lru_lock, flags); +	if (PageLRU(newpage)) +		del_page_from_lru_list(zone, newpage, page_lru(newpage)); +	__mem_cgroup_commit_charge(memcg, newpage, 1, pc, type); +	if (PageLRU(newpage)) +		add_page_to_lru_list(zone, newpage, page_lru(newpage)); +	spin_unlock_irqrestore(&zone->lru_lock, flags); +} +  #ifdef CONFIG_DEBUG_VM  static struct page_cgroup *lookup_page_cgroup_used(struct page *page)  {  |