________________________________ > From: Kirill Tkhai <[email protected]> > Sent: Tuesday, July 21, 2020 4:52 PM > To: Valeriy Vdovin <[email protected]>; [email protected] > > <[email protected]> > Cc: Valeriy Vdovin <[email protected]> > Subject: Re: [PATCH RHEL7 v20 06/14] ve/cgroup: unmark ve-root cgroups at > container stop > > On 25.06.2020 17:29, Valeriy Vdovin wrote: > > Signed-off-by: Valeriy Vdovin <[email protected]> > > Reviewed-by: Kirill Tkhai <[email protected]> > > --- > > include/linux/cgroup.h | 1 + > > kernel/cgroup.c | 38 ++++++++++++++++++++++++++++++++++++++ > > kernel/ve/ve.c | 2 ++ > > 3 files changed, 41 insertions(+) > > > > diff --git a/include/linux/cgroup.h b/include/linux/cgroup.h > > index ac60aaed..6e2c206 100644 > > --- a/include/linux/cgroup.h > > +++ b/include/linux/cgroup.h > > @@ -671,6 +671,7 @@ int cgroup_task_count(const struct cgroup *cgrp); > > > > #ifdef CONFIG_VE > > void cgroup_mark_ve_roots(struct ve_struct *ve); > > +void cgroup_unmark_ve_roots(struct ve_struct *ve); > > #endif > > > > /* > > diff --git a/kernel/cgroup.c b/kernel/cgroup.c > > index ce576c5..6e3871a 100644 > > --- a/kernel/cgroup.c > > +++ b/kernel/cgroup.c > > @@ -637,6 +637,31 @@ static struct css_set *find_css_set( > > } > > > > /* > > + * Walk each cgroup link of a given css_set and find a cgroup that > > + * is the child of cgroupfs_root in argument. > > + */ > > +static struct cgroup *css_cgroup_from_root(struct css_set *css_set, > > + struct cgroupfs_root *root) > > +{ > > + struct cgroup *res = NULL; > > + struct cg_cgroup_link *link; > > + > > + BUG_ON(!mutex_is_locked(&cgroup_mutex)); > > + read_lock(&css_set_lock); > > + > > + list_for_each_entry(link, &css_set->cg_links, cg_link_list) { > > + struct cgroup *c = link->cgrp; > > + if (c->root == root) { > > + res = c; > > + break; > > + } > > + } > > + read_unlock(&css_set_lock); > > + BUG_ON(!res); > > + return res; > > +} > > + > > +/* > > * Return the cgroup for "task" from the given hierarchy. Must be > > * called with cgroup_mutex held. > > */ > > @@ -4329,6 +4354,19 @@ void cgroup_mark_ve_roots(struct ve_struct *ve) > > mutex_unlock(&cgroup_mutex); > > } > > > > +void cgroup_unmark_ve_roots(struct ve_struct *ve) > > +{ > > + struct cgroup *cgrp; > > + struct cgroupfs_root *root; > > + > > + mutex_lock(&cgroup_mutex); > > + for_each_active_root(root) { > > + cgrp = css_cgroup_from_root(ve->root_css_set, root); > > + clear_bit(CGRP_VE_ROOT, &cgrp->flags); > > + } > > + mutex_unlock(&cgroup_mutex); > > +} > > + > > struct cgroup *cgroup_get_ve_root(struct cgroup *cgrp) > > { > > struct cgroup *ve_root = NULL; > > diff --git a/kernel/ve/ve.c b/kernel/ve/ve.c > > index 73cfee6..711050c 100644 > > --- a/kernel/ve/ve.c > > +++ b/kernel/ve/ve.c > > @@ -623,6 +623,8 @@ void ve_exit_ns(struct pid_namespace *pid_ns) > > if (!ve->ve_ns || ve->ve_ns->pid_ns != pid_ns) > > return; > > > > + cgroup_unmark_ve_roots(ve); > > Is there a problem that ve workqueue works will run after we unmark roots? > Maybe we should call this cgroup_unmark_ve_roots() after ve_workqueue_stop()?
When a cgroup gets empty it's decided to which workqueue it should be put to await for release. Thus when we unmark ve root, we prevent any new empty cgroups from entering this workqueue. After that we are safe to stop the workqueue by waiting for all the current jobs to complete. > > + > > ve_workqueue_stop(ve); > > > > /* > >
_______________________________________________ Devel mailing list [email protected] https://lists.openvz.org/mailman/listinfo/devel
