Loading kernel/trace/kmemtrace.c +67 −61 Original line number Diff line number Diff line Loading @@ -6,15 +6,16 @@ * Copyright (C) 2008 Frederic Weisbecker <fweisbec@gmail.com> */ #include <linux/dcache.h> #include <linux/tracepoint.h> #include <linux/seq_file.h> #include <linux/debugfs.h> #include <linux/dcache.h> #include <linux/fs.h> #include <linux/seq_file.h> #include <linux/tracepoint.h> #include <trace/kmemtrace.h> #include "trace.h" #include "trace_output.h" #include "trace.h" /* Select an alternative, minimalistic output than the original one */ #define TRACE_KMEM_OPT_MINIMAL 0x1 Loading @@ -41,13 +42,14 @@ static inline void kmemtrace_alloc(enum kmemtrace_type_id type_id, gfp_t gfp_flags, int node) { struct ring_buffer_event *event; struct kmemtrace_alloc_entry *entry; struct trace_array *tr = kmemtrace_array; struct kmemtrace_alloc_entry *entry; struct ring_buffer_event *event; event = ring_buffer_lock_reserve(tr->buffer, sizeof(*entry)); if (!event) return; entry = ring_buffer_event_data(event); tracing_generic_entry_update(&entry->ent, 0, 0); Loading @@ -69,9 +71,9 @@ static inline void kmemtrace_free(enum kmemtrace_type_id type_id, unsigned long call_site, const void *ptr) { struct ring_buffer_event *event; struct kmemtrace_free_entry *entry; struct trace_array *tr = kmemtrace_array; struct kmemtrace_free_entry *entry; struct ring_buffer_event *event; event = ring_buffer_lock_reserve(tr->buffer, sizeof(*entry)); if (!event) Loading Loading @@ -236,13 +238,14 @@ static enum print_line_t kmemtrace_print_alloc_user(struct trace_iterator *iter, struct kmemtrace_alloc_entry *entry) { struct kmemtrace_user_event_alloc *ev_alloc; struct trace_seq *s = &iter->seq; struct kmemtrace_user_event *ev; struct kmemtrace_user_event_alloc *ev_alloc; ev = trace_seq_reserve(s, sizeof(*ev)); if (!ev) return TRACE_TYPE_PARTIAL_LINE; ev->event_id = KMEMTRACE_USER_ALLOC; ev->type_id = entry->type_id; ev->event_size = sizeof(*ev) + sizeof(*ev_alloc); Loading @@ -254,6 +257,7 @@ kmemtrace_print_alloc_user(struct trace_iterator *iter, ev_alloc = trace_seq_reserve(s, sizeof(*ev_alloc)); if (!ev_alloc) return TRACE_TYPE_PARTIAL_LINE; ev_alloc->bytes_req = entry->bytes_req; ev_alloc->bytes_alloc = entry->bytes_alloc; ev_alloc->gfp_flags = entry->gfp_flags; Loading @@ -272,6 +276,7 @@ kmemtrace_print_free_user(struct trace_iterator *iter, ev = trace_seq_reserve(s, sizeof(*ev)); if (!ev) return TRACE_TYPE_PARTIAL_LINE; ev->event_id = KMEMTRACE_USER_FREE; ev->type_id = entry->type_id; ev->event_size = sizeof(*ev); Loading Loading @@ -415,6 +420,7 @@ static enum print_line_t kmemtrace_print_line(struct trace_iterator *iter) switch (entry->type) { case TRACE_KMEM_ALLOC: { struct kmemtrace_alloc_entry *field; trace_assign_type(field, entry); if (kmem_tracer_flags.val & TRACE_KMEM_OPT_MINIMAL) return kmemtrace_print_alloc_compress(iter, field); Loading @@ -424,6 +430,7 @@ static enum print_line_t kmemtrace_print_line(struct trace_iterator *iter) case TRACE_KMEM_FREE: { struct kmemtrace_free_entry *field; trace_assign_type(field, entry); if (kmem_tracer_flags.val & TRACE_KMEM_OPT_MINIMAL) return kmemtrace_print_free_compress(iter, field); Loading Loading @@ -454,5 +461,4 @@ static int __init init_kmem_tracer(void) { return register_tracer(&kmem_tracer); } device_initcall(init_kmem_tracer); Loading
kernel/trace/kmemtrace.c +67 −61 Original line number Diff line number Diff line Loading @@ -6,15 +6,16 @@ * Copyright (C) 2008 Frederic Weisbecker <fweisbec@gmail.com> */ #include <linux/dcache.h> #include <linux/tracepoint.h> #include <linux/seq_file.h> #include <linux/debugfs.h> #include <linux/dcache.h> #include <linux/fs.h> #include <linux/seq_file.h> #include <linux/tracepoint.h> #include <trace/kmemtrace.h> #include "trace.h" #include "trace_output.h" #include "trace.h" /* Select an alternative, minimalistic output than the original one */ #define TRACE_KMEM_OPT_MINIMAL 0x1 Loading @@ -41,13 +42,14 @@ static inline void kmemtrace_alloc(enum kmemtrace_type_id type_id, gfp_t gfp_flags, int node) { struct ring_buffer_event *event; struct kmemtrace_alloc_entry *entry; struct trace_array *tr = kmemtrace_array; struct kmemtrace_alloc_entry *entry; struct ring_buffer_event *event; event = ring_buffer_lock_reserve(tr->buffer, sizeof(*entry)); if (!event) return; entry = ring_buffer_event_data(event); tracing_generic_entry_update(&entry->ent, 0, 0); Loading @@ -69,9 +71,9 @@ static inline void kmemtrace_free(enum kmemtrace_type_id type_id, unsigned long call_site, const void *ptr) { struct ring_buffer_event *event; struct kmemtrace_free_entry *entry; struct trace_array *tr = kmemtrace_array; struct kmemtrace_free_entry *entry; struct ring_buffer_event *event; event = ring_buffer_lock_reserve(tr->buffer, sizeof(*entry)); if (!event) Loading Loading @@ -236,13 +238,14 @@ static enum print_line_t kmemtrace_print_alloc_user(struct trace_iterator *iter, struct kmemtrace_alloc_entry *entry) { struct kmemtrace_user_event_alloc *ev_alloc; struct trace_seq *s = &iter->seq; struct kmemtrace_user_event *ev; struct kmemtrace_user_event_alloc *ev_alloc; ev = trace_seq_reserve(s, sizeof(*ev)); if (!ev) return TRACE_TYPE_PARTIAL_LINE; ev->event_id = KMEMTRACE_USER_ALLOC; ev->type_id = entry->type_id; ev->event_size = sizeof(*ev) + sizeof(*ev_alloc); Loading @@ -254,6 +257,7 @@ kmemtrace_print_alloc_user(struct trace_iterator *iter, ev_alloc = trace_seq_reserve(s, sizeof(*ev_alloc)); if (!ev_alloc) return TRACE_TYPE_PARTIAL_LINE; ev_alloc->bytes_req = entry->bytes_req; ev_alloc->bytes_alloc = entry->bytes_alloc; ev_alloc->gfp_flags = entry->gfp_flags; Loading @@ -272,6 +276,7 @@ kmemtrace_print_free_user(struct trace_iterator *iter, ev = trace_seq_reserve(s, sizeof(*ev)); if (!ev) return TRACE_TYPE_PARTIAL_LINE; ev->event_id = KMEMTRACE_USER_FREE; ev->type_id = entry->type_id; ev->event_size = sizeof(*ev); Loading Loading @@ -415,6 +420,7 @@ static enum print_line_t kmemtrace_print_line(struct trace_iterator *iter) switch (entry->type) { case TRACE_KMEM_ALLOC: { struct kmemtrace_alloc_entry *field; trace_assign_type(field, entry); if (kmem_tracer_flags.val & TRACE_KMEM_OPT_MINIMAL) return kmemtrace_print_alloc_compress(iter, field); Loading @@ -424,6 +430,7 @@ static enum print_line_t kmemtrace_print_line(struct trace_iterator *iter) case TRACE_KMEM_FREE: { struct kmemtrace_free_entry *field; trace_assign_type(field, entry); if (kmem_tracer_flags.val & TRACE_KMEM_OPT_MINIMAL) return kmemtrace_print_free_compress(iter, field); Loading Loading @@ -454,5 +461,4 @@ static int __init init_kmem_tracer(void) { return register_tracer(&kmem_tracer); } device_initcall(init_kmem_tracer);