From: Francis Giraldeau <[email protected]>

Signed-off-by: Francis Giraldeau <[email protected]>
Signed-off-by: Mathieu Desnoyers <[email protected]>
---
 Makefile                  |   3 +-
 lttng-abi.c               |   3 +
 lttng-abi.h               |   2 +
 lttng-context-callstack.c | 334 ++++++++++++++++++++++++++++++++++++++++++++++
 lttng-events.h            |   5 +
 5 files changed, 346 insertions(+), 1 deletion(-)
 create mode 100644 lttng-context-callstack.c

diff --git a/Makefile b/Makefile
index f65e463..2ea8ae4 100644
--- a/Makefile
+++ b/Makefile
@@ -51,7 +51,8 @@ ifneq ($(KERNELRELEASE),)
                        lttng-context-vtid.o lttng-context-ppid.o \
                        lttng-context-vppid.o lttng-context-cpu-id.o \
                        lttng-context-interruptible.o \
-                       lttng-context-need-reschedule.o lttng-calibrate.o \
+                       lttng-context-need-reschedule.o \
+                       lttng-context-callstack.o lttng-calibrate.o \
                        lttng-context-hostname.o wrapper/random.o \
                        probes/lttng.o wrapper/trace-clock.o \
                        wrapper/page_alloc.o \
diff --git a/lttng-abi.c b/lttng-abi.c
index f2b207c..7795370 100644
--- a/lttng-abi.c
+++ b/lttng-abi.c
@@ -246,6 +246,9 @@ long lttng_abi_add_context(struct file *file,
                return lttng_add_preemptible_to_ctx(ctx);
        case LTTNG_KERNEL_CONTEXT_MIGRATABLE:
                return lttng_add_migratable_to_ctx(ctx);
+       case LTTNG_KERNEL_CONTEXT_CALLSTACK_KERNEL:
+       case LTTNG_KERNEL_CONTEXT_CALLSTACK_USER:
+               return lttng_add_callstack_to_ctx(ctx, context_param->ctx);
        default:
                return -EINVAL;
        }
diff --git a/lttng-abi.h b/lttng-abi.h
index dac8658..c8c4e29 100644
--- a/lttng-abi.h
+++ b/lttng-abi.h
@@ -146,6 +146,8 @@ enum lttng_kernel_context_type {
        LTTNG_KERNEL_CONTEXT_PREEMPTIBLE        = 13,
        LTTNG_KERNEL_CONTEXT_NEED_RESCHEDULE    = 14,
        LTTNG_KERNEL_CONTEXT_MIGRATABLE         = 15,
+       LTTNG_KERNEL_CONTEXT_CALLSTACK_KERNEL   = 16,
+       LTTNG_KERNEL_CONTEXT_CALLSTACK_USER     = 17,
 };
 
 struct lttng_kernel_perf_counter_ctx {
diff --git a/lttng-context-callstack.c b/lttng-context-callstack.c
new file mode 100644
index 0000000..8c3ee3d
--- /dev/null
+++ b/lttng-context-callstack.c
@@ -0,0 +1,334 @@
+/*
+ * lttng-context-callstack.c
+ *
+ * LTTng callstack event context.
+ *
+ * Copyright (C) 2014 Mathieu Desnoyers <[email protected]>
+ * Copyright (C) 2014 Francis Giraldeau <[email protected]>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; only
+ * version 2.1 of the License.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with this library; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ *
+ * The callstack context can be added to any kernel
+ * event. It records either the kernel or the userspace callstack, up to a
+ * max depth. The context is a CTF sequence, such that it uses only the space
+ * required for the number of callstack entries.
+ *
+ * It allocates callstack buffers per-CPU up to 4 interrupt nesting. This
+ * nesting limit is the same as defined in the ring buffer. It therefore uses a
+ * fixed amount of memory, proportional to the number of CPUs:
+ *
+ *   size = cpus * nest * depth * sizeof(unsigned long)
+ *
+ * Which is about 800 bytes per-CPUs on 64-bit host and a depth of 25. The
+ * allocation is done at the initialization to avoid memory allocation
+ * overhead while tracing, using a shallow stack.
+ *
+ * The kernel callstack is recovered using save_stack_trace(), and the
+ * userspace callstack uses save_stack_trace_user(). They rely on frame
+ * pointers. These are usually available for the kernel, but the compiler
+ * option -fomit-frame-pointer frequently used in popular Linux distributions
+ * may cause the userspace callstack to be unreliable, and is a known
+ * limitation of this approach. If frame pointers are not available, it
+ * produces no error, but the callstack will be empty. We still provide the
+ * feature, because it works well for runtime environments having frame
+ * pointers. In the future, unwind support and/or last branch record may
+ * provide a solution to this problem.
+ *
+ * The symbol name resolution is left to the trace reader.
+ */
+
+#include <linux/module.h>
+#include <linux/slab.h>
+#include <linux/sched.h>
+#include <linux/utsname.h>
+#include <linux/stacktrace.h>
+#include <linux/spinlock.h>
+#include "lttng-events.h"
+#include "wrapper/ringbuffer/backend.h"
+#include "wrapper/ringbuffer/frontend.h"
+#include "wrapper/vmalloc.h"
+#include "lttng-tracer.h"
+
+#define MAX_ENTRIES 25 /* BUG: saving more than 30 entries causes trace 
corruption */
+
+struct lttng_cs {
+       struct stack_trace items[RING_BUFFER_MAX_NESTING];
+};
+
+struct field_data {
+       int mode;
+       struct lttng_cs __percpu *cs_percpu;
+};
+
+struct lttng_cs_type {
+       const char *name;
+       const char *save_func_name;
+       void (*save_func)(struct stack_trace *trace);
+};
+
+enum lttng_cs_ctx_modes {
+       CALLSTACK_KERNEL = 0,
+       CALLSTACK_USER = 1,
+};
+
+static struct lttng_cs_type cs_types[] = {
+       {
+         .name           = "callstack_kernel",
+         .save_func_name = "save_stack_trace",
+         .save_func      = NULL,
+       },
+       {
+         .name           = "callstack_user",
+         .save_func_name = "save_stack_trace_user",
+         .save_func      = NULL,
+       },
+};
+
+static
+int init_type(int mode)
+{
+       unsigned long func;
+
+       if (cs_types[mode].save_func)
+               return 0;
+       func = kallsyms_lookup_funcptr(cs_types[mode].save_func_name);
+       if (!func) {
+               printk(KERN_WARNING "LTTng: symbol lookup failed: %s\n",
+                               cs_types[mode].save_func_name);
+               return -EINVAL;
+       }
+       cs_types[mode].save_func = (void *) func;
+       return 0;
+}
+
+static
+struct stack_trace *stack_trace_context(struct lttng_ctx_field *field,
+                                       struct lib_ring_buffer_ctx *ctx)
+{
+       int nesting;
+       struct lttng_cs *cs;
+       struct field_data *fdata = field->private;
+
+       /*
+        * get_cpu() is not required, preemption is already
+        * disabled while event is written.
+        *
+        * max nesting is checked in lib_ring_buffer_get_cpu().
+        * Check it again as a safety net.
+        */
+       cs = per_cpu_ptr(fdata->cs_percpu, ctx->cpu);
+       nesting = per_cpu(lib_ring_buffer_nesting, ctx->cpu) - 1;
+       if (nesting >= RING_BUFFER_MAX_NESTING) {
+               return NULL;
+       }
+       return &cs->items[nesting];
+}
+
+/*
+ * In order to reserve the correct size, the callstack is computed. The
+ * resulting callstack is saved to be accessed in the record step.
+ */
+static
+size_t lttng_callstack_get_size(size_t offset, struct lttng_ctx_field *field,
+                               struct lib_ring_buffer_ctx *ctx,
+                               struct lttng_channel *chan)
+{
+       size_t size = 0;
+       struct stack_trace *trace;
+       struct field_data *fdata = field->private;
+
+       /* do not write data if no space is available */
+       trace = stack_trace_context(field, ctx);
+       if (!trace)
+               return 0;
+
+       /* reset stack trace, no need to clear memory */
+       trace->nr_entries = 0;
+
+       /* do the real work and reserve space */
+       cs_types[fdata->mode].save_func(trace);
+       size += lib_ring_buffer_align(offset, lttng_alignof(unsigned int));
+       size += sizeof(unsigned int);
+       size += lib_ring_buffer_align(offset, lttng_alignof(unsigned long));
+       size += sizeof(unsigned long) * trace->nr_entries;
+       return size;
+}
+
+static
+void lttng_callstack_record(struct lttng_ctx_field *field,
+                           struct lib_ring_buffer_ctx *ctx,
+                           struct lttng_channel *chan)
+{
+       struct stack_trace *trace = stack_trace_context(field, ctx);
+
+       if (!trace)
+               return;
+       lib_ring_buffer_align_ctx(ctx, lttng_alignof(unsigned int));
+       chan->ops->event_write(ctx, &trace->nr_entries, sizeof(unsigned int));
+       lib_ring_buffer_align_ctx(ctx, lttng_alignof(unsigned long));
+       chan->ops->event_write(ctx, trace->entries,
+                       sizeof(unsigned long) * trace->nr_entries);
+}
+
+static
+void field_data_free(struct field_data *fdata)
+{
+       int cpu, i;
+       struct lttng_cs *cs;
+
+       if (!fdata)
+               return;
+       for_each_possible_cpu(cpu) {
+               cs = per_cpu_ptr(fdata->cs_percpu, cpu);
+               for (i = 0; i < RING_BUFFER_MAX_NESTING; i++) {
+                       kfree(cs->items[i].entries);
+               }
+       }
+       free_percpu(fdata->cs_percpu);
+       kfree(fdata);
+}
+
+static
+struct field_data __percpu *field_data_create(unsigned int entries, int type)
+{
+       int cpu, i;
+       struct stack_trace *item;
+       struct lttng_cs *cs;
+       struct lttng_cs __percpu *cs_set;
+       struct field_data* fdata;
+
+       fdata = kzalloc(sizeof(unsigned long) * entries, GFP_KERNEL);
+       if (!fdata)
+               return NULL;
+       cs_set = alloc_percpu(struct lttng_cs);
+       if (!cs_set)
+               goto error_alloc;
+
+       fdata->cs_percpu = cs_set;
+       for_each_possible_cpu(cpu) {
+               cs = per_cpu_ptr(cs_set, cpu);
+               for (i = 0; i < RING_BUFFER_MAX_NESTING; i++) {
+                       item = &cs->items[i];
+                       item->entries = kzalloc(sizeof(unsigned long) * 
entries, GFP_KERNEL);
+                       if (!item->entries) {
+                               goto error_alloc;
+                       }
+                       item->max_entries = entries;
+               }
+       }
+       fdata->mode = type;
+       return fdata;
+
+error_alloc:
+       field_data_free(fdata);
+       return NULL;
+}
+
+static
+void lttng_callstack_destroy(struct lttng_ctx_field *field)
+{
+       struct field_data *fdata = field->private;
+
+       field_data_free(fdata);
+}
+
+static
+int __lttng_add_callstack_generic(struct lttng_ctx **ctx, int mode)
+{
+       const char *ctx_name = cs_types[mode].name;
+       struct lttng_ctx_field *field;
+       struct field_data *fdata;
+       int ret;
+
+       ret = init_type(mode);
+       if (ret)
+               return ret;
+       field = lttng_append_context(ctx);
+       if (!field)
+               return -ENOMEM;
+       if (lttng_find_context(*ctx, ctx_name)) {
+               printk("%s lttng_find_context failed\n", ctx_name);
+               ret = -EEXIST;
+               goto error_find;
+       }
+       fdata = field_data_create(MAX_ENTRIES, mode);
+       if (!fdata) {
+               ret = -ENOMEM;
+               goto error_create;
+       }
+
+       field->event_field.name = ctx_name;
+       field->event_field.type.atype = atype_sequence;
+       field->event_field.type.u.sequence.elem_type.atype = atype_integer;
+       field->event_field.type.u.sequence.elem_type.u.basic.integer.size = 
sizeof(unsigned long) * CHAR_BIT;
+       field->event_field.type.u.sequence.elem_type.u.basic.integer.alignment 
= lttng_alignof(long) * CHAR_BIT;
+       field->event_field.type.u.sequence.elem_type.u.basic.integer.signedness 
= lttng_is_signed_type(unsigned long);
+       
field->event_field.type.u.sequence.elem_type.u.basic.integer.reverse_byte_order 
= 0;
+       field->event_field.type.u.sequence.elem_type.u.basic.integer.base = 16;
+       field->event_field.type.u.sequence.elem_type.u.basic.integer.encoding = 
lttng_encode_none;
+
+       field->event_field.type.u.sequence.length_type.atype = atype_integer;
+       field->event_field.type.u.sequence.length_type.u.basic.integer.size = 
sizeof(unsigned int) * CHAR_BIT;
+       
field->event_field.type.u.sequence.length_type.u.basic.integer.alignment = 
lttng_alignof(unsigned int) * CHAR_BIT;
+       
field->event_field.type.u.sequence.length_type.u.basic.integer.signedness = 
lttng_is_signed_type(unsigned int);
+       
field->event_field.type.u.sequence.length_type.u.basic.integer.reverse_byte_order
 = 0;
+       field->event_field.type.u.sequence.length_type.u.basic.integer.base = 
10;
+       field->event_field.type.u.sequence.length_type.u.basic.integer.encoding 
= lttng_encode_none;
+
+       field->get_size_arg = lttng_callstack_get_size;
+       field->record = lttng_callstack_record;
+       field->private = fdata;
+       field->destroy = lttng_callstack_destroy;
+       wrapper_vmalloc_sync_all();
+       printk("lttng add-context %s\n", ctx_name);
+       return 0;
+
+error_create:
+       field_data_free(fdata);
+error_find:
+       lttng_remove_context_field(ctx, field);
+       return ret;
+}
+
+/**
+ *     lttng_add_callstack_to_ctx - add callstack event context
+ *
+ *     @ctx: the lttng_ctx pointer to initialize
+ *     @type: the context type
+ *
+ *     Supported callstack type supported:
+ *     LTTNG_KERNEL_CONTEXT_CALLSTACK_KERNEL
+ *             Records the callstack of the kernel
+ *     LTTNG_KERNEL_CONTEXT_CALLSTACK_USER
+ *             Records the callstack of the userspace program (from the kernel)
+ *
+ * Return 0 for success, or error code.
+ */
+int lttng_add_callstack_to_ctx(struct lttng_ctx **ctx, int type)
+{
+       switch (type) {
+       case LTTNG_KERNEL_CONTEXT_CALLSTACK_KERNEL:
+               return __lttng_add_callstack_generic(ctx, CALLSTACK_KERNEL);
+       case LTTNG_KERNEL_CONTEXT_CALLSTACK_USER:
+               return __lttng_add_callstack_generic(ctx, CALLSTACK_USER);
+       default:
+               return -EINVAL;
+       }
+}
+EXPORT_SYMBOL_GPL(lttng_add_callstack_to_ctx);
+
+MODULE_LICENSE("GPL and additional rights");
+MODULE_AUTHOR("Francis Giraldeau");
+MODULE_DESCRIPTION("Linux Trace Toolkit Callstack Support");
diff --git a/lttng-events.h b/lttng-events.h
index 7a618be..05dd8de 100644
--- a/lttng-events.h
+++ b/lttng-events.h
@@ -217,6 +217,8 @@ struct lttng_ctx_field {
                struct lttng_perf_counter_field *perf_counter;
        } u;
        void (*destroy)(struct lttng_ctx_field *field);
+       /* private data to keep state between get_size and record */
+       void *private;
 };
 
 struct lttng_ctx {
@@ -705,6 +707,9 @@ int lttng_add_migratable_to_ctx(struct lttng_ctx **ctx)
        return -ENOSYS;
 }
 #endif
+
+int lttng_add_callstack_to_ctx(struct lttng_ctx **ctx, int type);
+
 #if defined(CONFIG_PERF_EVENTS) && (LINUX_VERSION_CODE >= 
KERNEL_VERSION(2,6,33))
 int lttng_add_perf_counter_to_ctx(uint32_t type,
                                  uint64_t config,
-- 
2.1.4

_______________________________________________
lttng-dev mailing list
[email protected]
https://lists.lttng.org/cgi-bin/mailman/listinfo/lttng-dev

Reply via email to