[Devel] [PATCH rh7 v4 4/9] mm/mem_cgroup_iter: Always assign iter->last_visited under rcu

Kirill Tkhai ktkhai at virtuozzo.com
Fri Feb 26 12:12:25 MSK 2021


On 24.02.2021 21:55, Konstantin Khorenko wrote:
> It's quite strange to have rcu section in mem_cgroup_iter(),
> but do not use rcu_dereference/rcu_assign for pointers being defended.
> 
> We plan to access/assign '->last_visited' during iterator invalidation,
> so we'll need the protection there anyway.
> 
> https://jira.sw.ru/browse/PSBM-123655
> 
> Signed-off-by: Konstantin Khorenko <khorenko at virtuozzo.com>
> ---
>  mm/memcontrol.c | 20 ++++++++++++++------
>  1 file changed, 14 insertions(+), 6 deletions(-)
> 
> diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> index 8040f09425bf..d0251d27de00 100644
> --- a/mm/memcontrol.c
> +++ b/mm/memcontrol.c
> @@ -194,8 +194,18 @@ struct mem_cgroup_reclaim_iter {
>  	/*
>  	 * last scanned hierarchy member. Valid only if last_dead_count
>  	 * matches memcg->dead_count of the hierarchy root group.
> +	 *
> +	 * Memory pointed by 'last_visited' is freed not earlier than
> +	 * one rcu period after we accessed it:
> +	 *   cgroup_offline_fn()
> +	 *    offline_css()
> +	 *    list_del_rcu()
> +	 *    dput()
> +	 *    ...
> +	 *     cgroup_diput()
> +	 *      call_rcu(&cgrp->rcu_head, cgroup_free_rcu)
>  	 */
> -	struct mem_cgroup *last_visited;
> +	struct mem_cgroup __rcu *last_visited;
>  	unsigned long last_dead_count;
>  
>  	/* scan generation, increased every round-trip */
> @@ -1591,8 +1601,7 @@ mem_cgroup_iter_load(struct mem_cgroup_reclaim_iter *iter,
>  	 */
>  	*sequence = atomic_read(&root->dead_count);
>  	if (iter->last_dead_count == *sequence) {
> -		smp_rmb();
> -		position = iter->last_visited;
> +		position = rcu_dereference(iter->last_visited);
>  
>  		/*
>  		 * We cannot take a reference to root because we might race
> @@ -1620,8 +1629,7 @@ static void mem_cgroup_iter_update(struct mem_cgroup_reclaim_iter *iter,
>  	 * don't lose destruction events in between.  We could have
>  	 * raced with the destruction of @new_position after all.
>  	 */
> -	iter->last_visited = new_position;
> -	smp_wmb();

We can't remove barriers in this patch, this makes the patch wrong.
We should remove barriers somewhere in [9/9].

> +	rcu_assign_pointer(iter->last_visited, new_position);
>  	iter->last_dead_count = sequence;
>  
>  	/* root reference counting symmetric to mem_cgroup_iter_load */
> @@ -1681,7 +1689,7 @@ struct mem_cgroup *mem_cgroup_iter(struct mem_cgroup *root,
>  			mz = mem_cgroup_zoneinfo(root, nid, zid);
>  			iter = &mz->reclaim_iter[reclaim->priority];
>  			if (prev && reclaim->generation != iter->generation) {
> -				iter->last_visited = NULL;
> +				rcu_assign_pointer(iter->last_visited, NULL);
>  				goto out_unlock;
>  			}
>  
> 



More information about the Devel mailing list