On Wed 04-04-18 10:18:50, Johannes Weiner wrote:
> On Wed, Apr 04, 2018 at 10:08:55AM -0400, Johannes Weiner wrote:
> > On Sat, Mar 24, 2018 at 09:08:59AM -0700, Tejun Heo wrote:
> > > @@ -91,6 +91,9 @@ struct mem_cgroup_stat_cpu {
> > >   unsigned long events[MEMCG_NR_EVENTS];
> > >   unsigned long nr_page_events;
> > >   unsigned long targets[MEM_CGROUP_NTARGETS];
> > > +
> > > + /* for cgroup rstat delta calculation */
> > > + unsigned long last_events[MEMCG_NR_EVENTS];
> > >  };
> > >  
> > >  struct mem_cgroup_reclaim_iter {
> > > @@ -233,7 +236,11 @@ struct mem_cgroup {
> > >  
> > >   struct mem_cgroup_stat_cpu __percpu *stat_cpu;
> > >   atomic_long_t           stat[MEMCG_NR_STAT];
> > > - atomic_long_t           events[MEMCG_NR_EVENTS];
> > > +
> > > + /* events is managed by cgroup rstat */
> > > + unsigned long long      events[MEMCG_NR_EVENTS];        /* local */
> > > + unsigned long long      tree_events[MEMCG_NR_EVENTS];   /* subtree */
> > > + unsigned long long      pending_events[MEMCG_NR_EVENTS];/* propagation 
> > > */
> > 
> > The lazy updates are neat, but I'm a little concerned at the memory
> > footprint. On a 64-cpu machine for example, this adds close to 9000
> > words to struct mem_cgroup. And we really only need the accuracy for
> > the 4 cgroup items in memory.events, not all VM events and stats.
> > 
> > Why not restrict the patch to those? It would also get rid of the
> > weird sharing between VM and cgroup enums.
> 
> In fact, I wonder if we need per-cpuness for MEMCG_LOW, MEMCG_HIGH
> etc. in the first place. They describe super high-level reclaim and
> OOM events, so they're not nearly as hot as other VM events and
> stats. We could probably just have a per-memcg array of atomics.

Agreed!

-- 
Michal Hocko
SUSE Labs

Reply via email to