On Mon, Apr 29, 2019 at 11:31:26AM -0400, Liang, Kan wrote: > > > On 4/29/2019 11:12 AM, Mark Rutland wrote: > > On Mon, Apr 29, 2019 at 07:44:03AM -0700, [email protected] wrote: > > > From: Kan Liang <[email protected]> > > > > > > A fast path will be introduced in the following patches to speed up the > > > cgroup events sched in, which only needs a simpler filter_match(). > > > > > > Add filter_match() as a parameter for pinned/flexible_sched_in(). > > > > > > No functional change. > > > > I suspect that the cost you're trying to avoid is pmu_filter_match() > > iterating over the entire group, which arm systems rely upon for correct > > behaviour on big.LITTLE systems. > > > > Is that the case? > > No. In X86, we don't use pmu_filter_match(). The fast path still keeps this > filter. > perf_cgroup_match() is the one I want to avoid.
Understood; sorry for the silly question, and thanks for confirming! :) Thanks, Mark.

