[Devel] [PATCH RHEL7 v20 06/14] ve/cgroup: unmark ve-root cgroups at container stop
Kirill Tkhai
ktkhai at virtuozzo.com
Mon Jul 27 13:18:11 MSK 2020
On 27.07.2020 12:10, Valeriy Vdovin wrote:
>
>
> ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
>> From: Kirill Tkhai <ktkhai at virtuozzo.com>
>> Sent: Tuesday, July 21, 2020 4:52 PM
>> To: Valeriy Vdovin <Valeriy.Vdovin at virtuozzo.com>; devel at openvz.org > <devel at openvz.org>
>> Cc: Valeriy Vdovin <valeriy.vdovin at virtuozz.com>
>> Subject: Re: [PATCH RHEL7 v20 06/14] ve/cgroup: unmark ve-root cgroups at container stop
>>
>> On 25.06.2020 17:29, Valeriy Vdovin wrote:
>> > Signed-off-by: Valeriy Vdovin <valeriy.vdovin at virtuozzo.com>
>> > Reviewed-by: Kirill Tkhai <ktkhai at virtuozzo.com>
>> > ---
>> > include/linux/cgroup.h | 1 +
>> > kernel/cgroup.c | 38 ++++++++++++++++++++++++++++++++++++++
>> > kernel/ve/ve.c | 2 ++
>> > 3 files changed, 41 insertions(+)
>> >
>> > diff --git a/include/linux/cgroup.h b/include/linux/cgroup.h
>> > index ac60aaed..6e2c206 100644
>> > --- a/include/linux/cgroup.h
>> > +++ b/include/linux/cgroup.h
>> > @@ -671,6 +671,7 @@ int cgroup_task_count(const struct cgroup *cgrp);
>> >
>> > #ifdef CONFIG_VE
>> > void cgroup_mark_ve_roots(struct ve_struct *ve);
>> > +void cgroup_unmark_ve_roots(struct ve_struct *ve);
>> > #endif
>> >
>> > /*
>> > diff --git a/kernel/cgroup.c b/kernel/cgroup.c
>> > index ce576c5..6e3871a 100644
>> > --- a/kernel/cgroup.c
>> > +++ b/kernel/cgroup.c
>> > @@ -637,6 +637,31 @@ static struct css_set *find_css_set(
>> > }
>> >
>> > /*
>> > + * Walk each cgroup link of a given css_set and find a cgroup that
>> > + * is the child of cgroupfs_root in argument.
>> > + */
>> > +static struct cgroup *css_cgroup_from_root(struct css_set *css_set,
>> > + struct cgroupfs_root *root)
>> > +{
>> > + struct cgroup *res = NULL;
>> > + struct cg_cgroup_link *link;
>> > +
>> > + BUG_ON(!mutex_is_locked(&cgroup_mutex));
>> > + read_lock(&css_set_lock);
>> > +
>> > + list_for_each_entry(link, &css_set->cg_links, cg_link_list) {
>> > + struct cgroup *c = link->cgrp;
>> > + if (c->root == root) {
>> > + res = c;
>> > + break;
>> > + }
>> > + }
>> > + read_unlock(&css_set_lock);
>> > + BUG_ON(!res);
>> > + return res;
>> > +}
>> > +
>> > +/*
>> > * Return the cgroup for "task" from the given hierarchy. Must be
>> > * called with cgroup_mutex held.
>> > */
>> > @@ -4329,6 +4354,19 @@ void cgroup_mark_ve_roots(struct ve_struct *ve)
>> > mutex_unlock(&cgroup_mutex);
>> > }
>> >
>> > +void cgroup_unmark_ve_roots(struct ve_struct *ve)
>> > +{
>> > + struct cgroup *cgrp;
>> > + struct cgroupfs_root *root;
>> > +
>> > + mutex_lock(&cgroup_mutex);
>> > + for_each_active_root(root) {
>> > + cgrp = css_cgroup_from_root(ve->root_css_set, root);
>> > + clear_bit(CGRP_VE_ROOT, &cgrp->flags);
>> > + }
>> > + mutex_unlock(&cgroup_mutex);
>> > +}
>> > +
>> > struct cgroup *cgroup_get_ve_root(struct cgroup *cgrp)
>> > {
>> > struct cgroup *ve_root = NULL;
>> > diff --git a/kernel/ve/ve.c b/kernel/ve/ve.c
>> > index 73cfee6..711050c 100644
>> > --- a/kernel/ve/ve.c
>> > +++ b/kernel/ve/ve.c
>> > @@ -623,6 +623,8 @@ void ve_exit_ns(struct pid_namespace *pid_ns)
>> > if (!ve->ve_ns || ve->ve_ns->pid_ns != pid_ns)
>> > return;
>> >
>> > + cgroup_unmark_ve_roots(ve);
>>
>> Is there a problem that ve workqueue works will run after we unmark roots?
>> Maybe we should call this cgroup_unmark_ve_roots() after ve_workqueue_stop()?
>
> When a cgroup gets empty it's decided to which workqueue it should be put to await for
> release. Thus when we unmark ve root, we prevent any new empty cgroups from entering this
> workqueue. After that we are safe to stop the workqueue by waiting for all the current jobs to
> complete.
But is it OK that workqueue is executing its works after root is cleared? Won't there some
security leak? If it's safe, why?
>
>> > +
>> > ve_workqueue_stop(ve);
>> >
>> > /*
>> >
>
More information about the Devel
mailing list