[Devel] Re: [PATCH 23/29] memory controller memory accounting v7

Peter Zijlstra a.p.zijlstra at chello.nl
Wed Sep 12 13:56:50 PDT 2007


> From: Balbir Singh <balbir at linux.vnet.ibm.com>

>  void page_assign_page_cgroup(struct page *page, struct page_cgroup *pc)
>  {
> -	page->page_cgroup = (unsigned long)pc;
> +	int locked;
> +
> +	/*
> +	 * While resetting the page_cgroup we might not hold the
> +	 * page_cgroup lock. free_hot_cold_page() is an example
> +	 * of such a scenario
> +	 */
> +	if (pc)
> +		VM_BUG_ON(!page_cgroup_locked(page));
> +	locked = (page->page_cgroup & PAGE_CGROUP_LOCK);
> +	page->page_cgroup = ((unsigned long)pc | locked);
>  }

This looks a bit odd, why not write:

 locked = page_cgroup_locked(page);
 if (pc)
   VM_BUG_ON(!locked)

> +/*
> + * Charge the memory controller for page usage.
> + * Return
> + * 0 if the charge was successful
> + * < 0 if the cgroup is over its limit
> + */
> +int mem_cgroup_charge(struct page *page, struct mm_struct *mm)
> +{
> +	struct mem_cgroup *mem;
> +	struct page_cgroup *pc, *race_pc;
> +
> +	/*
> +	 * Should page_cgroup's go to their own slab?
> +	 * One could optimize the performance of the charging routine
> +	 * by saving a bit in the page_flags and using it as a lock
> +	 * to see if the cgroup page already has a page_cgroup associated
> +	 * with it
> +	 */
> +	lock_page_cgroup(page);
> +	pc = page_get_page_cgroup(page);
> +	/*
> +	 * The page_cgroup exists and the page has already been accounted
> +	 */
> +	if (pc) {
> +		atomic_inc(&pc->ref_cnt);
> +		goto done;
> +	}
> +
> +	unlock_page_cgroup(page);
> +
> +	pc = kzalloc(sizeof(struct page_cgroup), GFP_KERNEL);
> +	if (pc == NULL)
> +		goto err;
> +
> +	rcu_read_lock();
> +	/*
> +	 * We always charge the cgroup the mm_struct belongs to
> +	 * the mm_struct's mem_cgroup changes on task migration if the
> +	 * thread group leader migrates. It's possible that mm is not
> +	 * set, if so charge the init_mm (happens for pagecache usage).
> +	 */
> +	if (!mm)
> +		mm = &init_mm;
> +
> +	mem = rcu_dereference(mm->mem_cgroup);
> +	/*
> +	 * For every charge from the cgroup, increment reference
> +	 * count
> +	 */
> +	css_get(&mem->css);
> +	rcu_read_unlock();
> +
> +	/*
> +	 * If we created the page_cgroup, we should free it on exceeding
> +	 * the cgroup limit.
> +	 */
> +	if (res_counter_charge(&mem->res, 1)) {
> +		css_put(&mem->css);
> +		goto free_pc;
> +	}
> +
> +	lock_page_cgroup(page);
> +	/*
> +	 * Check if somebody else beat us to allocating the page_cgroup
> +	 */
> +	race_pc = page_get_page_cgroup(page);
> +	if (race_pc) {
> +		kfree(pc);
> +		pc = race_pc;
> +		atomic_inc(&pc->ref_cnt);

This inc

> +		res_counter_uncharge(&mem->res, 1);
> +		css_put(&mem->css);
> +		goto done;
> +	}
> +
> +	atomic_set(&pc->ref_cnt, 1);

combined with this set make me wonder...

> +	pc->mem_cgroup = mem;
> +	pc->page = page;
> +	page_assign_page_cgroup(page, pc);
> +
> +done:
> +	unlock_page_cgroup(page);
> +	return 0;
> +free_pc:
> +	kfree(pc);
> +	return -ENOMEM;
> +err:
> +	unlock_page_cgroup(page);
> +	return -ENOMEM;
> +}



> @@ -2161,6 +2184,9 @@ static int do_anonymous_page(struct mm_s
>  	if (!page)
>  		goto oom;
>  
> +		if (mem_cgroup_charge(page, mm))
> +			goto oom_free_page;
> +
>  	entry = mk_pte(page, vma->vm_page_prot);
>  	entry = maybe_mkwrite(pte_mkdirty(entry), vma);
>  

whitespace damage




_______________________________________________
Containers mailing list
Containers at lists.linux-foundation.org
https://lists.linux-foundation.org/mailman/listinfo/containers




More information about the Devel mailing list