[Devel] [PATCH RHEL7 v20 06/14] ve/cgroup: unmark ve-root cgroups at container stop

Valeriy Vdovin Valeriy.Vdovin at virtuozzo.com
Mon Jul 27 12:10:14 MSK 2020



________________________________
> From: Kirill Tkhai <ktkhai at virtuozzo.com>
> Sent: Tuesday, July 21, 2020 4:52 PM
> To: Valeriy Vdovin <Valeriy.Vdovin at virtuozzo.com>; devel at openvz.org > <devel at openvz.org>
> Cc: Valeriy Vdovin <valeriy.vdovin at virtuozz.com>
> Subject: Re: [PATCH RHEL7 v20 06/14] ve/cgroup: unmark ve-root cgroups at container stop
>
> On 25.06.2020 17:29, Valeriy Vdovin wrote:
> > Signed-off-by: Valeriy Vdovin <valeriy.vdovin at virtuozzo.com>
> > Reviewed-by: Kirill Tkhai <ktkhai at virtuozzo.com>
> > ---
> >  include/linux/cgroup.h |  1 +
> >  kernel/cgroup.c        | 38 ++++++++++++++++++++++++++++++++++++++
> >  kernel/ve/ve.c         |  2 ++
> >  3 files changed, 41 insertions(+)
> >
> > diff --git a/include/linux/cgroup.h b/include/linux/cgroup.h
> > index ac60aaed..6e2c206 100644
> > --- a/include/linux/cgroup.h
> > +++ b/include/linux/cgroup.h
> > @@ -671,6 +671,7 @@ int cgroup_task_count(const struct cgroup *cgrp);
> >
> >  #ifdef CONFIG_VE
> >  void cgroup_mark_ve_roots(struct ve_struct *ve);
> > +void cgroup_unmark_ve_roots(struct ve_struct *ve);
> >  #endif
> >
> >  /*
> > diff --git a/kernel/cgroup.c b/kernel/cgroup.c
> > index ce576c5..6e3871a 100644
> > --- a/kernel/cgroup.c
> > +++ b/kernel/cgroup.c
> > @@ -637,6 +637,31 @@ static struct css_set *find_css_set(
> >  }
> >
> >  /*
> > + * Walk each cgroup link of a given css_set and find a cgroup that
> > + * is the child of cgroupfs_root in argument.
> > + */
> > +static struct cgroup *css_cgroup_from_root(struct css_set *css_set,
> > +                                         struct cgroupfs_root *root)
> > +{
> > +     struct cgroup *res = NULL;
> > +     struct cg_cgroup_link *link;
> > +
> > +     BUG_ON(!mutex_is_locked(&cgroup_mutex));
> > +     read_lock(&css_set_lock);
> > +
> > +     list_for_each_entry(link, &css_set->cg_links, cg_link_list) {
> > +             struct cgroup *c = link->cgrp;
> > +             if (c->root == root) {
> > +                     res = c;
> > +                     break;
> > +             }
> > +     }
> > +     read_unlock(&css_set_lock);
> > +     BUG_ON(!res);
> > +     return res;
> > +}
> > +
> > +/*
> >   * Return the cgroup for "task" from the given hierarchy. Must be
> >   * called with cgroup_mutex held.
> >   */
> > @@ -4329,6 +4354,19 @@ void cgroup_mark_ve_roots(struct ve_struct *ve)
> >        mutex_unlock(&cgroup_mutex);
> >  }
> >
> > +void cgroup_unmark_ve_roots(struct ve_struct *ve)
> > +{
> > +     struct cgroup *cgrp;
> > +     struct cgroupfs_root *root;
> > +
> > +     mutex_lock(&cgroup_mutex);
> > +     for_each_active_root(root) {
> > +             cgrp = css_cgroup_from_root(ve->root_css_set, root);
> > +             clear_bit(CGRP_VE_ROOT, &cgrp->flags);
> > +     }
> > +     mutex_unlock(&cgroup_mutex);
> > +}
> > +
> >  struct cgroup *cgroup_get_ve_root(struct cgroup *cgrp)
> >  {
> >        struct cgroup *ve_root = NULL;
> > diff --git a/kernel/ve/ve.c b/kernel/ve/ve.c
> > index 73cfee6..711050c 100644
> > --- a/kernel/ve/ve.c
> > +++ b/kernel/ve/ve.c
> > @@ -623,6 +623,8 @@ void ve_exit_ns(struct pid_namespace *pid_ns)
> >        if (!ve->ve_ns || ve->ve_ns->pid_ns != pid_ns)
> >                return;
> >
> > +     cgroup_unmark_ve_roots(ve);
>
> Is there a problem that ve workqueue works will run after we unmark roots?
> Maybe we should call this cgroup_unmark_ve_roots() after ve_workqueue_stop()?

When a cgroup gets empty it's decided to which workqueue it should be put to await for
release. Thus when we unmark ve root, we prevent any new empty cgroups from entering this
workqueue. After that we are safe to stop the workqueue by waiting for all the current jobs to
complete.

> > +
> >        ve_workqueue_stop(ve);
> >
> >        /*
> >

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openvz.org/pipermail/devel/attachments/20200727/a182d17c/attachment.html>


More information about the Devel mailing list