This patch exposes the branch trace buffer to users for sampling. There are measurements where it is very useful to couple the instruction address with some path information, e.g, basic block profiling.
On recent Intel processors, the branch stack is implemented using the LBR registers. LBR was already used to fixup PEBS. This patch still allows PEBS fixups with LBR and also exposes LBR to applications. There is a new PERF_SAMPLE_BRANCH_STACK sample type. It creates a sample in the buffer which has the following layout: { u64 nr; { u64 from, to, flags } lbr[nr]; } && PERF_SAMPLE_BRANCH_STACK }; Refer to include/linux/perf_event.h to figure out the layout ordering information. LBR is configured by default to record ALL taken branches. On some processors, it is possible to filter the type of branches. This will be supported in a subsequent patch. On other processors, the sample type is allowed but will generate a sample where nr=0 as is the case with other sampling types. Signed-off-by: Stephane Eranian <eran...@google.com> -- arch/x86/kernel/cpu/perf_event_intel.c | 13 +++++++++++++ arch/x86/kernel/cpu/perf_event_intel_ds.c | 5 +++++ include/linux/perf_event.h | 8 +++++++- kernel/perf_event.c | 25 +++++++++++++++++++++++++ 4 files changed, 50 insertions(+), 1 deletion(-) diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c index f168b40..6b8aa7d 100644 --- a/arch/x86/kernel/cpu/perf_event_intel.c +++ b/arch/x86/kernel/cpu/perf_event_intel.c @@ -562,8 +562,13 @@ static void intel_pmu_disable_event(struct perf_event *event) x86_pmu_disable_event(event); + /* + * PEBS implies LBR + */ if (unlikely(event->attr.precise)) intel_pmu_pebs_disable(event); + else if (unlikely(event->attr.sample_type & PERF_SAMPLE_BRANCH_STACK)) + intel_pmu_lbr_disable(event); } static void intel_pmu_enable_fixed(struct hw_perf_event *hwc) @@ -614,8 +619,13 @@ static void intel_pmu_enable_event(struct perf_event *event) return; } + /* + * PEBS implies LBR + */ if (unlikely(event->attr.precise)) intel_pmu_pebs_enable(event); + else if (unlikely(event->attr.sample_type & PERF_SAMPLE_BRANCH_STACK)) + intel_pmu_lbr_enable(event); __x86_pmu_enable_event(hwc); } @@ -710,6 +720,9 @@ again: data.period = event->hw.last_period; + if (event->attr.sample_type & PERF_SAMPLE_BRANCH_STACK) + data.br_stack = &cpuc->lbr_stack; + if (perf_event_overflow(event, 1, &data, regs)) x86_pmu_stop(event); } diff --git a/arch/x86/kernel/cpu/perf_event_intel_ds.c b/arch/x86/kernel/cpu/perf_event_intel_ds.c index 2fea362..ed62d35 100644 --- a/arch/x86/kernel/cpu/perf_event_intel_ds.c +++ b/arch/x86/kernel/cpu/perf_event_intel_ds.c @@ -505,6 +505,8 @@ static void intel_pmu_drain_pebs_core(struct pt_regs *iregs) data.raw = &raw; } + if (event->attr.sample_type & PERF_SAMPLE_BRANCH_STACK) + data.br_stack = &cpuc->lbr_stack; /* * We use the interrupt regs as a base because the PEBS record * does not contain a full regs set, specifically it seems to @@ -591,6 +593,9 @@ static void intel_pmu_drain_pebs_nhm(struct pt_regs *iregs) data.raw = &raw; } + if (event->attr.sample_type & PERF_SAMPLE_BRANCH_STACK) + data.br_stack = &cpuc->lbr_stack; + /* * See the comment in intel_pmu_drain_pebs_core() */ diff --git a/include/linux/perf_event.h b/include/linux/perf_event.h index 6e96cc8..3a8288f 100644 --- a/include/linux/perf_event.h +++ b/include/linux/perf_event.h @@ -125,8 +125,9 @@ enum perf_event_sample_format { PERF_SAMPLE_PERIOD = 1U << 8, PERF_SAMPLE_STREAM_ID = 1U << 9, PERF_SAMPLE_RAW = 1U << 10, + PERF_SAMPLE_BRANCH_STACK = 1U << 11, - PERF_SAMPLE_MAX = 1U << 11, /* non-ABI */ + PERF_SAMPLE_MAX = 1U << 12, /* non-ABI */ }; /* @@ -415,6 +416,9 @@ enum perf_event_type { * * { u32 size; * char data[size];}&& PERF_SAMPLE_RAW + * + * { u64 nr; + * { u64 from, to, flags } lbr[nr]; } && PERF_SAMPLE_BRANCH_STACK * }; */ PERF_RECORD_SAMPLE = 9, @@ -819,6 +823,7 @@ struct perf_sample_data { u64 period; struct perf_callchain_entry *callchain; struct perf_raw_record *raw; + struct perf_branch_stack *br_stack; }; static inline @@ -826,6 +831,7 @@ void perf_sample_data_init(struct perf_sample_data *data, u64 addr) { data->addr = addr; data->raw = NULL; + data->br_stack = NULL; } extern void perf_output_sample(struct perf_output_handle *handle, diff --git a/kernel/perf_event.c b/kernel/perf_event.c index 63fbce1..8143e77 100644 --- a/kernel/perf_event.c +++ b/kernel/perf_event.c @@ -3211,6 +3211,22 @@ void perf_output_sample(struct perf_output_handle *handle, perf_output_put(handle, raw); } } + + if (sample_type & PERF_SAMPLE_BRANCH_STACK) { + if (data->br_stack) { + size_t size; + + size = data->br_stack->nr + * sizeof(struct perf_branch_entry); + + perf_output_put(handle, data->br_stack->nr); + perf_output_copy(handle, data->br_stack->entries, size); + } else { + u64 nr = 0; + perf_output_put(handle, nr); + } + } + } void perf_prepare_sample(struct perf_event_header *header, @@ -3298,6 +3314,15 @@ void perf_prepare_sample(struct perf_event_header *header, WARN_ON_ONCE(size & (sizeof(u64)-1)); header->size += size; } + + if (sample_type & PERF_SAMPLE_BRANCH_STACK) { + int size = sizeof(u64); + if (data->br_stack) { + size += data->br_stack->nr + * sizeof(struct perf_branch_entry); + } + header->size += size; + } } static void perf_event_output(struct perf_event *event, int nmi, ------------------------------------------------------------------------------ Download Intel® Parallel Studio Eval Try the new software tools for yourself. Speed compiling, find bugs proactively, and fine-tune applications for parallel performance. See why Intel Parallel Studio got high marks during beta. http://p.sf.net/sfu/intel-sw-dev _______________________________________________ perfmon2-devel mailing list perfmon2-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/perfmon2-devel