Merge branch 'perf-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip
* 'perf-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip: (163 commits) tracing: Fix compile issue for trace_sched_wakeup.c [S390] hardirq: remove pointless header file includes [IA64] Move local_softirq_pending() definition perf, powerpc: Fix power_pmu_event_init to not use event->ctx ftrace: Remove recursion between recordmcount and scripts/mod/empty jump_label: Add COND_STMT(), reducer wrappery perf: Optimize sw events perf: Use jump_labels to optimize the scheduler hooks jump_label: Add atomic_t interface jump_label: Use more consistent naming perf, hw_breakpoint: Fix crash in hw_breakpoint creation perf: Find task before event alloc perf: Fix task refcount bugs perf: Fix group moving irq_work: Add generic hardirq context callbacks perf_events: Fix transaction recovery in group_sched_in() perf_events: Fix bogus AMD64 generic TLB events perf_events: Fix bogus context time tracking tracing: Remove parent recording in latency tracer graph options tracing: Use one prologue for the preempt irqs off tracer function tracers ...
This commit is contained in:
@@ -49,6 +49,11 @@ config HAVE_SYSCALL_TRACEPOINTS
|
||||
help
|
||||
See Documentation/trace/ftrace-design.txt
|
||||
|
||||
config HAVE_C_RECORDMCOUNT
|
||||
bool
|
||||
help
|
||||
C version of recordmcount available?
|
||||
|
||||
config TRACER_MAX_TRACE
|
||||
bool
|
||||
|
||||
|
@@ -884,10 +884,8 @@ enum {
|
||||
FTRACE_ENABLE_CALLS = (1 << 0),
|
||||
FTRACE_DISABLE_CALLS = (1 << 1),
|
||||
FTRACE_UPDATE_TRACE_FUNC = (1 << 2),
|
||||
FTRACE_ENABLE_MCOUNT = (1 << 3),
|
||||
FTRACE_DISABLE_MCOUNT = (1 << 4),
|
||||
FTRACE_START_FUNC_RET = (1 << 5),
|
||||
FTRACE_STOP_FUNC_RET = (1 << 6),
|
||||
FTRACE_START_FUNC_RET = (1 << 3),
|
||||
FTRACE_STOP_FUNC_RET = (1 << 4),
|
||||
};
|
||||
|
||||
static int ftrace_filtered;
|
||||
@@ -1226,8 +1224,6 @@ static void ftrace_shutdown(int command)
|
||||
|
||||
static void ftrace_startup_sysctl(void)
|
||||
{
|
||||
int command = FTRACE_ENABLE_MCOUNT;
|
||||
|
||||
if (unlikely(ftrace_disabled))
|
||||
return;
|
||||
|
||||
@@ -1235,23 +1231,17 @@ static void ftrace_startup_sysctl(void)
|
||||
saved_ftrace_func = NULL;
|
||||
/* ftrace_start_up is true if we want ftrace running */
|
||||
if (ftrace_start_up)
|
||||
command |= FTRACE_ENABLE_CALLS;
|
||||
|
||||
ftrace_run_update_code(command);
|
||||
ftrace_run_update_code(FTRACE_ENABLE_CALLS);
|
||||
}
|
||||
|
||||
static void ftrace_shutdown_sysctl(void)
|
||||
{
|
||||
int command = FTRACE_DISABLE_MCOUNT;
|
||||
|
||||
if (unlikely(ftrace_disabled))
|
||||
return;
|
||||
|
||||
/* ftrace_start_up is true if ftrace is running */
|
||||
if (ftrace_start_up)
|
||||
command |= FTRACE_DISABLE_CALLS;
|
||||
|
||||
ftrace_run_update_code(command);
|
||||
ftrace_run_update_code(FTRACE_DISABLE_CALLS);
|
||||
}
|
||||
|
||||
static cycle_t ftrace_update_time;
|
||||
@@ -1368,24 +1358,29 @@ enum {
|
||||
#define FTRACE_BUFF_MAX (KSYM_SYMBOL_LEN+4) /* room for wildcards */
|
||||
|
||||
struct ftrace_iterator {
|
||||
struct ftrace_page *pg;
|
||||
int hidx;
|
||||
int idx;
|
||||
unsigned flags;
|
||||
struct trace_parser parser;
|
||||
loff_t pos;
|
||||
loff_t func_pos;
|
||||
struct ftrace_page *pg;
|
||||
struct dyn_ftrace *func;
|
||||
struct ftrace_func_probe *probe;
|
||||
struct trace_parser parser;
|
||||
int hidx;
|
||||
int idx;
|
||||
unsigned flags;
|
||||
};
|
||||
|
||||
static void *
|
||||
t_hash_next(struct seq_file *m, void *v, loff_t *pos)
|
||||
t_hash_next(struct seq_file *m, loff_t *pos)
|
||||
{
|
||||
struct ftrace_iterator *iter = m->private;
|
||||
struct hlist_node *hnd = v;
|
||||
struct hlist_node *hnd = NULL;
|
||||
struct hlist_head *hhd;
|
||||
|
||||
WARN_ON(!(iter->flags & FTRACE_ITER_HASH));
|
||||
|
||||
(*pos)++;
|
||||
iter->pos = *pos;
|
||||
|
||||
if (iter->probe)
|
||||
hnd = &iter->probe->node;
|
||||
retry:
|
||||
if (iter->hidx >= FTRACE_FUNC_HASHSIZE)
|
||||
return NULL;
|
||||
@@ -1408,7 +1403,12 @@ t_hash_next(struct seq_file *m, void *v, loff_t *pos)
|
||||
}
|
||||
}
|
||||
|
||||
return hnd;
|
||||
if (WARN_ON_ONCE(!hnd))
|
||||
return NULL;
|
||||
|
||||
iter->probe = hlist_entry(hnd, struct ftrace_func_probe, node);
|
||||
|
||||
return iter;
|
||||
}
|
||||
|
||||
static void *t_hash_start(struct seq_file *m, loff_t *pos)
|
||||
@@ -1417,26 +1417,32 @@ static void *t_hash_start(struct seq_file *m, loff_t *pos)
|
||||
void *p = NULL;
|
||||
loff_t l;
|
||||
|
||||
if (!(iter->flags & FTRACE_ITER_HASH))
|
||||
*pos = 0;
|
||||
|
||||
iter->flags |= FTRACE_ITER_HASH;
|
||||
if (iter->func_pos > *pos)
|
||||
return NULL;
|
||||
|
||||
iter->hidx = 0;
|
||||
for (l = 0; l <= *pos; ) {
|
||||
p = t_hash_next(m, p, &l);
|
||||
for (l = 0; l <= (*pos - iter->func_pos); ) {
|
||||
p = t_hash_next(m, &l);
|
||||
if (!p)
|
||||
break;
|
||||
}
|
||||
return p;
|
||||
if (!p)
|
||||
return NULL;
|
||||
|
||||
/* Only set this if we have an item */
|
||||
iter->flags |= FTRACE_ITER_HASH;
|
||||
|
||||
return iter;
|
||||
}
|
||||
|
||||
static int t_hash_show(struct seq_file *m, void *v)
|
||||
static int
|
||||
t_hash_show(struct seq_file *m, struct ftrace_iterator *iter)
|
||||
{
|
||||
struct ftrace_func_probe *rec;
|
||||
struct hlist_node *hnd = v;
|
||||
|
||||
rec = hlist_entry(hnd, struct ftrace_func_probe, node);
|
||||
rec = iter->probe;
|
||||
if (WARN_ON_ONCE(!rec))
|
||||
return -EIO;
|
||||
|
||||
if (rec->ops->print)
|
||||
return rec->ops->print(m, rec->ip, rec->ops, rec->data);
|
||||
@@ -1457,12 +1463,13 @@ t_next(struct seq_file *m, void *v, loff_t *pos)
|
||||
struct dyn_ftrace *rec = NULL;
|
||||
|
||||
if (iter->flags & FTRACE_ITER_HASH)
|
||||
return t_hash_next(m, v, pos);
|
||||
return t_hash_next(m, pos);
|
||||
|
||||
(*pos)++;
|
||||
iter->pos = *pos;
|
||||
|
||||
if (iter->flags & FTRACE_ITER_PRINTALL)
|
||||
return NULL;
|
||||
return t_hash_start(m, pos);
|
||||
|
||||
retry:
|
||||
if (iter->idx >= iter->pg->index) {
|
||||
@@ -1491,7 +1498,20 @@ t_next(struct seq_file *m, void *v, loff_t *pos)
|
||||
}
|
||||
}
|
||||
|
||||
return rec;
|
||||
if (!rec)
|
||||
return t_hash_start(m, pos);
|
||||
|
||||
iter->func_pos = *pos;
|
||||
iter->func = rec;
|
||||
|
||||
return iter;
|
||||
}
|
||||
|
||||
static void reset_iter_read(struct ftrace_iterator *iter)
|
||||
{
|
||||
iter->pos = 0;
|
||||
iter->func_pos = 0;
|
||||
iter->flags &= ~(FTRACE_ITER_PRINTALL & FTRACE_ITER_HASH);
|
||||
}
|
||||
|
||||
static void *t_start(struct seq_file *m, loff_t *pos)
|
||||
@@ -1501,6 +1521,12 @@ static void *t_start(struct seq_file *m, loff_t *pos)
|
||||
loff_t l;
|
||||
|
||||
mutex_lock(&ftrace_lock);
|
||||
/*
|
||||
* If an lseek was done, then reset and start from beginning.
|
||||
*/
|
||||
if (*pos < iter->pos)
|
||||
reset_iter_read(iter);
|
||||
|
||||
/*
|
||||
* For set_ftrace_filter reading, if we have the filter
|
||||
* off, we can short cut and just print out that all
|
||||
@@ -1518,6 +1544,11 @@ static void *t_start(struct seq_file *m, loff_t *pos)
|
||||
if (iter->flags & FTRACE_ITER_HASH)
|
||||
return t_hash_start(m, pos);
|
||||
|
||||
/*
|
||||
* Unfortunately, we need to restart at ftrace_pages_start
|
||||
* every time we let go of the ftrace_mutex. This is because
|
||||
* those pointers can change without the lock.
|
||||
*/
|
||||
iter->pg = ftrace_pages_start;
|
||||
iter->idx = 0;
|
||||
for (l = 0; l <= *pos; ) {
|
||||
@@ -1526,10 +1557,14 @@ static void *t_start(struct seq_file *m, loff_t *pos)
|
||||
break;
|
||||
}
|
||||
|
||||
if (!p && iter->flags & FTRACE_ITER_FILTER)
|
||||
return t_hash_start(m, pos);
|
||||
if (!p) {
|
||||
if (iter->flags & FTRACE_ITER_FILTER)
|
||||
return t_hash_start(m, pos);
|
||||
|
||||
return p;
|
||||
return NULL;
|
||||
}
|
||||
|
||||
return iter;
|
||||
}
|
||||
|
||||
static void t_stop(struct seq_file *m, void *p)
|
||||
@@ -1540,16 +1575,18 @@ static void t_stop(struct seq_file *m, void *p)
|
||||
static int t_show(struct seq_file *m, void *v)
|
||||
{
|
||||
struct ftrace_iterator *iter = m->private;
|
||||
struct dyn_ftrace *rec = v;
|
||||
struct dyn_ftrace *rec;
|
||||
|
||||
if (iter->flags & FTRACE_ITER_HASH)
|
||||
return t_hash_show(m, v);
|
||||
return t_hash_show(m, iter);
|
||||
|
||||
if (iter->flags & FTRACE_ITER_PRINTALL) {
|
||||
seq_printf(m, "#### all functions enabled ####\n");
|
||||
return 0;
|
||||
}
|
||||
|
||||
rec = iter->func;
|
||||
|
||||
if (!rec)
|
||||
return 0;
|
||||
|
||||
@@ -1601,8 +1638,8 @@ ftrace_failures_open(struct inode *inode, struct file *file)
|
||||
|
||||
ret = ftrace_avail_open(inode, file);
|
||||
if (!ret) {
|
||||
m = (struct seq_file *)file->private_data;
|
||||
iter = (struct ftrace_iterator *)m->private;
|
||||
m = file->private_data;
|
||||
iter = m->private;
|
||||
iter->flags = FTRACE_ITER_FAILURES;
|
||||
}
|
||||
|
||||
@@ -2418,7 +2455,7 @@ static const struct file_operations ftrace_filter_fops = {
|
||||
.open = ftrace_filter_open,
|
||||
.read = seq_read,
|
||||
.write = ftrace_filter_write,
|
||||
.llseek = no_llseek,
|
||||
.llseek = ftrace_regex_lseek,
|
||||
.release = ftrace_filter_release,
|
||||
};
|
||||
|
||||
|
@@ -2606,6 +2606,19 @@ void ring_buffer_record_enable_cpu(struct ring_buffer *buffer, int cpu)
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(ring_buffer_record_enable_cpu);
|
||||
|
||||
/*
|
||||
* The total entries in the ring buffer is the running counter
|
||||
* of entries entered into the ring buffer, minus the sum of
|
||||
* the entries read from the ring buffer and the number of
|
||||
* entries that were overwritten.
|
||||
*/
|
||||
static inline unsigned long
|
||||
rb_num_of_entries(struct ring_buffer_per_cpu *cpu_buffer)
|
||||
{
|
||||
return local_read(&cpu_buffer->entries) -
|
||||
(local_read(&cpu_buffer->overrun) + cpu_buffer->read);
|
||||
}
|
||||
|
||||
/**
|
||||
* ring_buffer_entries_cpu - get the number of entries in a cpu buffer
|
||||
* @buffer: The ring buffer
|
||||
@@ -2614,16 +2627,13 @@ EXPORT_SYMBOL_GPL(ring_buffer_record_enable_cpu);
|
||||
unsigned long ring_buffer_entries_cpu(struct ring_buffer *buffer, int cpu)
|
||||
{
|
||||
struct ring_buffer_per_cpu *cpu_buffer;
|
||||
unsigned long ret;
|
||||
|
||||
if (!cpumask_test_cpu(cpu, buffer->cpumask))
|
||||
return 0;
|
||||
|
||||
cpu_buffer = buffer->buffers[cpu];
|
||||
ret = (local_read(&cpu_buffer->entries) - local_read(&cpu_buffer->overrun))
|
||||
- cpu_buffer->read;
|
||||
|
||||
return ret;
|
||||
return rb_num_of_entries(cpu_buffer);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(ring_buffer_entries_cpu);
|
||||
|
||||
@@ -2684,8 +2694,7 @@ unsigned long ring_buffer_entries(struct ring_buffer *buffer)
|
||||
/* if you care about this being correct, lock the buffer */
|
||||
for_each_buffer_cpu(buffer, cpu) {
|
||||
cpu_buffer = buffer->buffers[cpu];
|
||||
entries += (local_read(&cpu_buffer->entries) -
|
||||
local_read(&cpu_buffer->overrun)) - cpu_buffer->read;
|
||||
entries += rb_num_of_entries(cpu_buffer);
|
||||
}
|
||||
|
||||
return entries;
|
||||
|
@@ -2196,7 +2196,7 @@ int tracing_open_generic(struct inode *inode, struct file *filp)
|
||||
|
||||
static int tracing_release(struct inode *inode, struct file *file)
|
||||
{
|
||||
struct seq_file *m = (struct seq_file *)file->private_data;
|
||||
struct seq_file *m = file->private_data;
|
||||
struct trace_iterator *iter;
|
||||
int cpu;
|
||||
|
||||
|
@@ -343,6 +343,10 @@ void trace_function(struct trace_array *tr,
|
||||
unsigned long ip,
|
||||
unsigned long parent_ip,
|
||||
unsigned long flags, int pc);
|
||||
void trace_graph_function(struct trace_array *tr,
|
||||
unsigned long ip,
|
||||
unsigned long parent_ip,
|
||||
unsigned long flags, int pc);
|
||||
void trace_default_header(struct seq_file *m);
|
||||
void print_trace_header(struct seq_file *m, struct trace_iterator *iter);
|
||||
int trace_empty(struct trace_iterator *iter);
|
||||
|
@@ -9,7 +9,7 @@
|
||||
#include <linux/kprobes.h>
|
||||
#include "trace.h"
|
||||
|
||||
static char *perf_trace_buf[4];
|
||||
static char __percpu *perf_trace_buf[PERF_NR_CONTEXTS];
|
||||
|
||||
/*
|
||||
* Force it to be aligned to unsigned long to avoid misaligned accesses
|
||||
@@ -24,7 +24,7 @@ static int total_ref_count;
|
||||
static int perf_trace_event_init(struct ftrace_event_call *tp_event,
|
||||
struct perf_event *p_event)
|
||||
{
|
||||
struct hlist_head *list;
|
||||
struct hlist_head __percpu *list;
|
||||
int ret = -ENOMEM;
|
||||
int cpu;
|
||||
|
||||
@@ -42,11 +42,11 @@ static int perf_trace_event_init(struct ftrace_event_call *tp_event,
|
||||
tp_event->perf_events = list;
|
||||
|
||||
if (!total_ref_count) {
|
||||
char *buf;
|
||||
char __percpu *buf;
|
||||
int i;
|
||||
|
||||
for (i = 0; i < 4; i++) {
|
||||
buf = (char *)alloc_percpu(perf_trace_t);
|
||||
for (i = 0; i < PERF_NR_CONTEXTS; i++) {
|
||||
buf = (char __percpu *)alloc_percpu(perf_trace_t);
|
||||
if (!buf)
|
||||
goto fail;
|
||||
|
||||
@@ -65,7 +65,7 @@ fail:
|
||||
if (!total_ref_count) {
|
||||
int i;
|
||||
|
||||
for (i = 0; i < 4; i++) {
|
||||
for (i = 0; i < PERF_NR_CONTEXTS; i++) {
|
||||
free_percpu(perf_trace_buf[i]);
|
||||
perf_trace_buf[i] = NULL;
|
||||
}
|
||||
@@ -101,22 +101,26 @@ int perf_trace_init(struct perf_event *p_event)
|
||||
return ret;
|
||||
}
|
||||
|
||||
int perf_trace_enable(struct perf_event *p_event)
|
||||
int perf_trace_add(struct perf_event *p_event, int flags)
|
||||
{
|
||||
struct ftrace_event_call *tp_event = p_event->tp_event;
|
||||
struct hlist_head __percpu *pcpu_list;
|
||||
struct hlist_head *list;
|
||||
|
||||
list = tp_event->perf_events;
|
||||
if (WARN_ON_ONCE(!list))
|
||||
pcpu_list = tp_event->perf_events;
|
||||
if (WARN_ON_ONCE(!pcpu_list))
|
||||
return -EINVAL;
|
||||
|
||||
list = this_cpu_ptr(list);
|
||||
if (!(flags & PERF_EF_START))
|
||||
p_event->hw.state = PERF_HES_STOPPED;
|
||||
|
||||
list = this_cpu_ptr(pcpu_list);
|
||||
hlist_add_head_rcu(&p_event->hlist_entry, list);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
void perf_trace_disable(struct perf_event *p_event)
|
||||
void perf_trace_del(struct perf_event *p_event, int flags)
|
||||
{
|
||||
hlist_del_rcu(&p_event->hlist_entry);
|
||||
}
|
||||
@@ -142,7 +146,7 @@ void perf_trace_destroy(struct perf_event *p_event)
|
||||
tp_event->perf_events = NULL;
|
||||
|
||||
if (!--total_ref_count) {
|
||||
for (i = 0; i < 4; i++) {
|
||||
for (i = 0; i < PERF_NR_CONTEXTS; i++) {
|
||||
free_percpu(perf_trace_buf[i]);
|
||||
perf_trace_buf[i] = NULL;
|
||||
}
|
||||
|
@@ -600,21 +600,29 @@ out:
|
||||
|
||||
enum {
|
||||
FORMAT_HEADER = 1,
|
||||
FORMAT_PRINTFMT = 2,
|
||||
FORMAT_FIELD_SEPERATOR = 2,
|
||||
FORMAT_PRINTFMT = 3,
|
||||
};
|
||||
|
||||
static void *f_next(struct seq_file *m, void *v, loff_t *pos)
|
||||
{
|
||||
struct ftrace_event_call *call = m->private;
|
||||
struct ftrace_event_field *field;
|
||||
struct list_head *head;
|
||||
struct list_head *common_head = &ftrace_common_fields;
|
||||
struct list_head *head = trace_get_fields(call);
|
||||
|
||||
(*pos)++;
|
||||
|
||||
switch ((unsigned long)v) {
|
||||
case FORMAT_HEADER:
|
||||
head = &ftrace_common_fields;
|
||||
if (unlikely(list_empty(common_head)))
|
||||
return NULL;
|
||||
|
||||
field = list_entry(common_head->prev,
|
||||
struct ftrace_event_field, link);
|
||||
return field;
|
||||
|
||||
case FORMAT_FIELD_SEPERATOR:
|
||||
if (unlikely(list_empty(head)))
|
||||
return NULL;
|
||||
|
||||
@@ -626,31 +634,10 @@ static void *f_next(struct seq_file *m, void *v, loff_t *pos)
|
||||
return NULL;
|
||||
}
|
||||
|
||||
head = trace_get_fields(call);
|
||||
|
||||
/*
|
||||
* To separate common fields from event fields, the
|
||||
* LSB is set on the first event field. Clear it in case.
|
||||
*/
|
||||
v = (void *)((unsigned long)v & ~1L);
|
||||
|
||||
field = v;
|
||||
/*
|
||||
* If this is a common field, and at the end of the list, then
|
||||
* continue with main list.
|
||||
*/
|
||||
if (field->link.prev == &ftrace_common_fields) {
|
||||
if (unlikely(list_empty(head)))
|
||||
return NULL;
|
||||
field = list_entry(head->prev, struct ftrace_event_field, link);
|
||||
/* Set the LSB to notify f_show to print an extra newline */
|
||||
field = (struct ftrace_event_field *)
|
||||
((unsigned long)field | 1);
|
||||
return field;
|
||||
}
|
||||
|
||||
/* If we are done tell f_show to print the format */
|
||||
if (field->link.prev == head)
|
||||
if (field->link.prev == common_head)
|
||||
return (void *)FORMAT_FIELD_SEPERATOR;
|
||||
else if (field->link.prev == head)
|
||||
return (void *)FORMAT_PRINTFMT;
|
||||
|
||||
field = list_entry(field->link.prev, struct ftrace_event_field, link);
|
||||
@@ -688,22 +675,16 @@ static int f_show(struct seq_file *m, void *v)
|
||||
seq_printf(m, "format:\n");
|
||||
return 0;
|
||||
|
||||
case FORMAT_FIELD_SEPERATOR:
|
||||
seq_putc(m, '\n');
|
||||
return 0;
|
||||
|
||||
case FORMAT_PRINTFMT:
|
||||
seq_printf(m, "\nprint fmt: %s\n",
|
||||
call->print_fmt);
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* To separate common fields from event fields, the
|
||||
* LSB is set on the first event field. Clear it and
|
||||
* print a newline if it is set.
|
||||
*/
|
||||
if ((unsigned long)v & 1) {
|
||||
seq_putc(m, '\n');
|
||||
v = (void *)((unsigned long)v & ~1L);
|
||||
}
|
||||
|
||||
field = v;
|
||||
|
||||
/*
|
||||
|
@@ -15,15 +15,19 @@
|
||||
#include "trace.h"
|
||||
#include "trace_output.h"
|
||||
|
||||
/* When set, irq functions will be ignored */
|
||||
static int ftrace_graph_skip_irqs;
|
||||
|
||||
struct fgraph_cpu_data {
|
||||
pid_t last_pid;
|
||||
int depth;
|
||||
int depth_irq;
|
||||
int ignore;
|
||||
unsigned long enter_funcs[FTRACE_RETFUNC_DEPTH];
|
||||
};
|
||||
|
||||
struct fgraph_data {
|
||||
struct fgraph_cpu_data *cpu_data;
|
||||
struct fgraph_cpu_data __percpu *cpu_data;
|
||||
|
||||
/* Place to preserve last processed entry. */
|
||||
struct ftrace_graph_ent_entry ent;
|
||||
@@ -41,6 +45,7 @@ struct fgraph_data {
|
||||
#define TRACE_GRAPH_PRINT_PROC 0x8
|
||||
#define TRACE_GRAPH_PRINT_DURATION 0x10
|
||||
#define TRACE_GRAPH_PRINT_ABS_TIME 0x20
|
||||
#define TRACE_GRAPH_PRINT_IRQS 0x40
|
||||
|
||||
static struct tracer_opt trace_opts[] = {
|
||||
/* Display overruns? (for self-debug purpose) */
|
||||
@@ -55,13 +60,15 @@ static struct tracer_opt trace_opts[] = {
|
||||
{ TRACER_OPT(funcgraph-duration, TRACE_GRAPH_PRINT_DURATION) },
|
||||
/* Display absolute time of an entry */
|
||||
{ TRACER_OPT(funcgraph-abstime, TRACE_GRAPH_PRINT_ABS_TIME) },
|
||||
/* Display interrupts */
|
||||
{ TRACER_OPT(funcgraph-irqs, TRACE_GRAPH_PRINT_IRQS) },
|
||||
{ } /* Empty entry */
|
||||
};
|
||||
|
||||
static struct tracer_flags tracer_flags = {
|
||||
/* Don't display overruns and proc by default */
|
||||
.val = TRACE_GRAPH_PRINT_CPU | TRACE_GRAPH_PRINT_OVERHEAD |
|
||||
TRACE_GRAPH_PRINT_DURATION,
|
||||
TRACE_GRAPH_PRINT_DURATION | TRACE_GRAPH_PRINT_IRQS,
|
||||
.opts = trace_opts
|
||||
};
|
||||
|
||||
@@ -204,6 +211,14 @@ int __trace_graph_entry(struct trace_array *tr,
|
||||
return 1;
|
||||
}
|
||||
|
||||
static inline int ftrace_graph_ignore_irqs(void)
|
||||
{
|
||||
if (!ftrace_graph_skip_irqs)
|
||||
return 0;
|
||||
|
||||
return in_irq();
|
||||
}
|
||||
|
||||
int trace_graph_entry(struct ftrace_graph_ent *trace)
|
||||
{
|
||||
struct trace_array *tr = graph_array;
|
||||
@@ -218,7 +233,8 @@ int trace_graph_entry(struct ftrace_graph_ent *trace)
|
||||
return 0;
|
||||
|
||||
/* trace it when it is-nested-in or is a function enabled. */
|
||||
if (!(trace->depth || ftrace_graph_addr(trace->func)))
|
||||
if (!(trace->depth || ftrace_graph_addr(trace->func)) ||
|
||||
ftrace_graph_ignore_irqs())
|
||||
return 0;
|
||||
|
||||
local_irq_save(flags);
|
||||
@@ -246,6 +262,34 @@ int trace_graph_thresh_entry(struct ftrace_graph_ent *trace)
|
||||
return trace_graph_entry(trace);
|
||||
}
|
||||
|
||||
static void
|
||||
__trace_graph_function(struct trace_array *tr,
|
||||
unsigned long ip, unsigned long flags, int pc)
|
||||
{
|
||||
u64 time = trace_clock_local();
|
||||
struct ftrace_graph_ent ent = {
|
||||
.func = ip,
|
||||
.depth = 0,
|
||||
};
|
||||
struct ftrace_graph_ret ret = {
|
||||
.func = ip,
|
||||
.depth = 0,
|
||||
.calltime = time,
|
||||
.rettime = time,
|
||||
};
|
||||
|
||||
__trace_graph_entry(tr, &ent, flags, pc);
|
||||
__trace_graph_return(tr, &ret, flags, pc);
|
||||
}
|
||||
|
||||
void
|
||||
trace_graph_function(struct trace_array *tr,
|
||||
unsigned long ip, unsigned long parent_ip,
|
||||
unsigned long flags, int pc)
|
||||
{
|
||||
__trace_graph_function(tr, ip, flags, pc);
|
||||
}
|
||||
|
||||
void __trace_graph_return(struct trace_array *tr,
|
||||
struct ftrace_graph_ret *trace,
|
||||
unsigned long flags,
|
||||
@@ -649,8 +693,9 @@ trace_print_graph_duration(unsigned long long duration, struct trace_seq *s)
|
||||
|
||||
/* Print nsecs (we don't want to exceed 7 numbers) */
|
||||
if (len < 7) {
|
||||
snprintf(nsecs_str, min(sizeof(nsecs_str), 8UL - len), "%03lu",
|
||||
nsecs_rem);
|
||||
size_t slen = min_t(size_t, sizeof(nsecs_str), 8UL - len);
|
||||
|
||||
snprintf(nsecs_str, slen, "%03lu", nsecs_rem);
|
||||
ret = trace_seq_printf(s, ".%s", nsecs_str);
|
||||
if (!ret)
|
||||
return TRACE_TYPE_PARTIAL_LINE;
|
||||
@@ -855,6 +900,108 @@ print_graph_prologue(struct trace_iterator *iter, struct trace_seq *s,
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* Entry check for irq code
|
||||
*
|
||||
* returns 1 if
|
||||
* - we are inside irq code
|
||||
* - we just extered irq code
|
||||
*
|
||||
* retunns 0 if
|
||||
* - funcgraph-interrupts option is set
|
||||
* - we are not inside irq code
|
||||
*/
|
||||
static int
|
||||
check_irq_entry(struct trace_iterator *iter, u32 flags,
|
||||
unsigned long addr, int depth)
|
||||
{
|
||||
int cpu = iter->cpu;
|
||||
int *depth_irq;
|
||||
struct fgraph_data *data = iter->private;
|
||||
|
||||
/*
|
||||
* If we are either displaying irqs, or we got called as
|
||||
* a graph event and private data does not exist,
|
||||
* then we bypass the irq check.
|
||||
*/
|
||||
if ((flags & TRACE_GRAPH_PRINT_IRQS) ||
|
||||
(!data))
|
||||
return 0;
|
||||
|
||||
depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
|
||||
|
||||
/*
|
||||
* We are inside the irq code
|
||||
*/
|
||||
if (*depth_irq >= 0)
|
||||
return 1;
|
||||
|
||||
if ((addr < (unsigned long)__irqentry_text_start) ||
|
||||
(addr >= (unsigned long)__irqentry_text_end))
|
||||
return 0;
|
||||
|
||||
/*
|
||||
* We are entering irq code.
|
||||
*/
|
||||
*depth_irq = depth;
|
||||
return 1;
|
||||
}
|
||||
|
||||
/*
|
||||
* Return check for irq code
|
||||
*
|
||||
* returns 1 if
|
||||
* - we are inside irq code
|
||||
* - we just left irq code
|
||||
*
|
||||
* returns 0 if
|
||||
* - funcgraph-interrupts option is set
|
||||
* - we are not inside irq code
|
||||
*/
|
||||
static int
|
||||
check_irq_return(struct trace_iterator *iter, u32 flags, int depth)
|
||||
{
|
||||
int cpu = iter->cpu;
|
||||
int *depth_irq;
|
||||
struct fgraph_data *data = iter->private;
|
||||
|
||||
/*
|
||||
* If we are either displaying irqs, or we got called as
|
||||
* a graph event and private data does not exist,
|
||||
* then we bypass the irq check.
|
||||
*/
|
||||
if ((flags & TRACE_GRAPH_PRINT_IRQS) ||
|
||||
(!data))
|
||||
return 0;
|
||||
|
||||
depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
|
||||
|
||||
/*
|
||||
* We are not inside the irq code.
|
||||
*/
|
||||
if (*depth_irq == -1)
|
||||
return 0;
|
||||
|
||||
/*
|
||||
* We are inside the irq code, and this is returning entry.
|
||||
* Let's not trace it and clear the entry depth, since
|
||||
* we are out of irq code.
|
||||
*
|
||||
* This condition ensures that we 'leave the irq code' once
|
||||
* we are out of the entry depth. Thus protecting us from
|
||||
* the RETURN entry loss.
|
||||
*/
|
||||
if (*depth_irq >= depth) {
|
||||
*depth_irq = -1;
|
||||
return 1;
|
||||
}
|
||||
|
||||
/*
|
||||
* We are inside the irq code, and this is not the entry.
|
||||
*/
|
||||
return 1;
|
||||
}
|
||||
|
||||
static enum print_line_t
|
||||
print_graph_entry(struct ftrace_graph_ent_entry *field, struct trace_seq *s,
|
||||
struct trace_iterator *iter, u32 flags)
|
||||
@@ -865,6 +1012,9 @@ print_graph_entry(struct ftrace_graph_ent_entry *field, struct trace_seq *s,
|
||||
static enum print_line_t ret;
|
||||
int cpu = iter->cpu;
|
||||
|
||||
if (check_irq_entry(iter, flags, call->func, call->depth))
|
||||
return TRACE_TYPE_HANDLED;
|
||||
|
||||
if (print_graph_prologue(iter, s, TRACE_GRAPH_ENT, call->func, flags))
|
||||
return TRACE_TYPE_PARTIAL_LINE;
|
||||
|
||||
@@ -902,6 +1052,9 @@ print_graph_return(struct ftrace_graph_ret *trace, struct trace_seq *s,
|
||||
int ret;
|
||||
int i;
|
||||
|
||||
if (check_irq_return(iter, flags, trace->depth))
|
||||
return TRACE_TYPE_HANDLED;
|
||||
|
||||
if (data) {
|
||||
struct fgraph_cpu_data *cpu_data;
|
||||
int cpu = iter->cpu;
|
||||
@@ -1054,7 +1207,7 @@ print_graph_comment(struct trace_seq *s, struct trace_entry *ent,
|
||||
|
||||
|
||||
enum print_line_t
|
||||
print_graph_function_flags(struct trace_iterator *iter, u32 flags)
|
||||
__print_graph_function_flags(struct trace_iterator *iter, u32 flags)
|
||||
{
|
||||
struct ftrace_graph_ent_entry *field;
|
||||
struct fgraph_data *data = iter->private;
|
||||
@@ -1117,7 +1270,18 @@ print_graph_function_flags(struct trace_iterator *iter, u32 flags)
|
||||
static enum print_line_t
|
||||
print_graph_function(struct trace_iterator *iter)
|
||||
{
|
||||
return print_graph_function_flags(iter, tracer_flags.val);
|
||||
return __print_graph_function_flags(iter, tracer_flags.val);
|
||||
}
|
||||
|
||||
enum print_line_t print_graph_function_flags(struct trace_iterator *iter,
|
||||
u32 flags)
|
||||
{
|
||||
if (trace_flags & TRACE_ITER_LATENCY_FMT)
|
||||
flags |= TRACE_GRAPH_PRINT_DURATION;
|
||||
else
|
||||
flags |= TRACE_GRAPH_PRINT_ABS_TIME;
|
||||
|
||||
return __print_graph_function_flags(iter, flags);
|
||||
}
|
||||
|
||||
static enum print_line_t
|
||||
@@ -1149,7 +1313,7 @@ static void print_lat_header(struct seq_file *s, u32 flags)
|
||||
seq_printf(s, "#%.*s|||| / \n", size, spaces);
|
||||
}
|
||||
|
||||
void print_graph_headers_flags(struct seq_file *s, u32 flags)
|
||||
static void __print_graph_headers_flags(struct seq_file *s, u32 flags)
|
||||
{
|
||||
int lat = trace_flags & TRACE_ITER_LATENCY_FMT;
|
||||
|
||||
@@ -1190,6 +1354,23 @@ void print_graph_headers(struct seq_file *s)
|
||||
print_graph_headers_flags(s, tracer_flags.val);
|
||||
}
|
||||
|
||||
void print_graph_headers_flags(struct seq_file *s, u32 flags)
|
||||
{
|
||||
struct trace_iterator *iter = s->private;
|
||||
|
||||
if (trace_flags & TRACE_ITER_LATENCY_FMT) {
|
||||
/* print nothing if the buffers are empty */
|
||||
if (trace_empty(iter))
|
||||
return;
|
||||
|
||||
print_trace_header(s, iter);
|
||||
flags |= TRACE_GRAPH_PRINT_DURATION;
|
||||
} else
|
||||
flags |= TRACE_GRAPH_PRINT_ABS_TIME;
|
||||
|
||||
__print_graph_headers_flags(s, flags);
|
||||
}
|
||||
|
||||
void graph_trace_open(struct trace_iterator *iter)
|
||||
{
|
||||
/* pid and depth on the last trace processed */
|
||||
@@ -1210,9 +1391,12 @@ void graph_trace_open(struct trace_iterator *iter)
|
||||
pid_t *pid = &(per_cpu_ptr(data->cpu_data, cpu)->last_pid);
|
||||
int *depth = &(per_cpu_ptr(data->cpu_data, cpu)->depth);
|
||||
int *ignore = &(per_cpu_ptr(data->cpu_data, cpu)->ignore);
|
||||
int *depth_irq = &(per_cpu_ptr(data->cpu_data, cpu)->depth_irq);
|
||||
|
||||
*pid = -1;
|
||||
*depth = 0;
|
||||
*ignore = 0;
|
||||
*depth_irq = -1;
|
||||
}
|
||||
|
||||
iter->private = data;
|
||||
@@ -1235,6 +1419,14 @@ void graph_trace_close(struct trace_iterator *iter)
|
||||
}
|
||||
}
|
||||
|
||||
static int func_graph_set_flag(u32 old_flags, u32 bit, int set)
|
||||
{
|
||||
if (bit == TRACE_GRAPH_PRINT_IRQS)
|
||||
ftrace_graph_skip_irqs = !set;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct trace_event_functions graph_functions = {
|
||||
.trace = print_graph_function_event,
|
||||
};
|
||||
@@ -1261,6 +1453,7 @@ static struct tracer graph_trace __read_mostly = {
|
||||
.print_line = print_graph_function,
|
||||
.print_header = print_graph_headers,
|
||||
.flags = &tracer_flags,
|
||||
.set_flag = func_graph_set_flag,
|
||||
#ifdef CONFIG_FTRACE_SELFTEST
|
||||
.selftest = trace_selftest_startup_function_graph,
|
||||
#endif
|
||||
|
@@ -87,14 +87,22 @@ static __cacheline_aligned_in_smp unsigned long max_sequence;
|
||||
|
||||
#ifdef CONFIG_FUNCTION_TRACER
|
||||
/*
|
||||
* irqsoff uses its own tracer function to keep the overhead down:
|
||||
* Prologue for the preempt and irqs off function tracers.
|
||||
*
|
||||
* Returns 1 if it is OK to continue, and data->disabled is
|
||||
* incremented.
|
||||
* 0 if the trace is to be ignored, and data->disabled
|
||||
* is kept the same.
|
||||
*
|
||||
* Note, this function is also used outside this ifdef but
|
||||
* inside the #ifdef of the function graph tracer below.
|
||||
* This is OK, since the function graph tracer is
|
||||
* dependent on the function tracer.
|
||||
*/
|
||||
static void
|
||||
irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip)
|
||||
static int func_prolog_dec(struct trace_array *tr,
|
||||
struct trace_array_cpu **data,
|
||||
unsigned long *flags)
|
||||
{
|
||||
struct trace_array *tr = irqsoff_trace;
|
||||
struct trace_array_cpu *data;
|
||||
unsigned long flags;
|
||||
long disabled;
|
||||
int cpu;
|
||||
|
||||
@@ -106,18 +114,38 @@ irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip)
|
||||
*/
|
||||
cpu = raw_smp_processor_id();
|
||||
if (likely(!per_cpu(tracing_cpu, cpu)))
|
||||
return;
|
||||
return 0;
|
||||
|
||||
local_save_flags(flags);
|
||||
local_save_flags(*flags);
|
||||
/* slight chance to get a false positive on tracing_cpu */
|
||||
if (!irqs_disabled_flags(flags))
|
||||
return;
|
||||
if (!irqs_disabled_flags(*flags))
|
||||
return 0;
|
||||
|
||||
data = tr->data[cpu];
|
||||
disabled = atomic_inc_return(&data->disabled);
|
||||
*data = tr->data[cpu];
|
||||
disabled = atomic_inc_return(&(*data)->disabled);
|
||||
|
||||
if (likely(disabled == 1))
|
||||
trace_function(tr, ip, parent_ip, flags, preempt_count());
|
||||
return 1;
|
||||
|
||||
atomic_dec(&(*data)->disabled);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* irqsoff uses its own tracer function to keep the overhead down:
|
||||
*/
|
||||
static void
|
||||
irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip)
|
||||
{
|
||||
struct trace_array *tr = irqsoff_trace;
|
||||
struct trace_array_cpu *data;
|
||||
unsigned long flags;
|
||||
|
||||
if (!func_prolog_dec(tr, &data, &flags))
|
||||
return;
|
||||
|
||||
trace_function(tr, ip, parent_ip, flags, preempt_count());
|
||||
|
||||
atomic_dec(&data->disabled);
|
||||
}
|
||||
@@ -155,30 +183,16 @@ static int irqsoff_graph_entry(struct ftrace_graph_ent *trace)
|
||||
struct trace_array *tr = irqsoff_trace;
|
||||
struct trace_array_cpu *data;
|
||||
unsigned long flags;
|
||||
long disabled;
|
||||
int ret;
|
||||
int cpu;
|
||||
int pc;
|
||||
|
||||
cpu = raw_smp_processor_id();
|
||||
if (likely(!per_cpu(tracing_cpu, cpu)))
|
||||
if (!func_prolog_dec(tr, &data, &flags))
|
||||
return 0;
|
||||
|
||||
local_save_flags(flags);
|
||||
/* slight chance to get a false positive on tracing_cpu */
|
||||
if (!irqs_disabled_flags(flags))
|
||||
return 0;
|
||||
|
||||
data = tr->data[cpu];
|
||||
disabled = atomic_inc_return(&data->disabled);
|
||||
|
||||
if (likely(disabled == 1)) {
|
||||
pc = preempt_count();
|
||||
ret = __trace_graph_entry(tr, trace, flags, pc);
|
||||
} else
|
||||
ret = 0;
|
||||
|
||||
pc = preempt_count();
|
||||
ret = __trace_graph_entry(tr, trace, flags, pc);
|
||||
atomic_dec(&data->disabled);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -187,27 +201,13 @@ static void irqsoff_graph_return(struct ftrace_graph_ret *trace)
|
||||
struct trace_array *tr = irqsoff_trace;
|
||||
struct trace_array_cpu *data;
|
||||
unsigned long flags;
|
||||
long disabled;
|
||||
int cpu;
|
||||
int pc;
|
||||
|
||||
cpu = raw_smp_processor_id();
|
||||
if (likely(!per_cpu(tracing_cpu, cpu)))
|
||||
if (!func_prolog_dec(tr, &data, &flags))
|
||||
return;
|
||||
|
||||
local_save_flags(flags);
|
||||
/* slight chance to get a false positive on tracing_cpu */
|
||||
if (!irqs_disabled_flags(flags))
|
||||
return;
|
||||
|
||||
data = tr->data[cpu];
|
||||
disabled = atomic_inc_return(&data->disabled);
|
||||
|
||||
if (likely(disabled == 1)) {
|
||||
pc = preempt_count();
|
||||
__trace_graph_return(tr, trace, flags, pc);
|
||||
}
|
||||
|
||||
pc = preempt_count();
|
||||
__trace_graph_return(tr, trace, flags, pc);
|
||||
atomic_dec(&data->disabled);
|
||||
}
|
||||
|
||||
@@ -229,75 +229,33 @@ static void irqsoff_trace_close(struct trace_iterator *iter)
|
||||
|
||||
static enum print_line_t irqsoff_print_line(struct trace_iterator *iter)
|
||||
{
|
||||
u32 flags = GRAPH_TRACER_FLAGS;
|
||||
|
||||
if (trace_flags & TRACE_ITER_LATENCY_FMT)
|
||||
flags |= TRACE_GRAPH_PRINT_DURATION;
|
||||
else
|
||||
flags |= TRACE_GRAPH_PRINT_ABS_TIME;
|
||||
|
||||
/*
|
||||
* In graph mode call the graph tracer output function,
|
||||
* otherwise go with the TRACE_FN event handler
|
||||
*/
|
||||
if (is_graph())
|
||||
return print_graph_function_flags(iter, flags);
|
||||
return print_graph_function_flags(iter, GRAPH_TRACER_FLAGS);
|
||||
|
||||
return TRACE_TYPE_UNHANDLED;
|
||||
}
|
||||
|
||||
static void irqsoff_print_header(struct seq_file *s)
|
||||
{
|
||||
if (is_graph()) {
|
||||
struct trace_iterator *iter = s->private;
|
||||
u32 flags = GRAPH_TRACER_FLAGS;
|
||||
|
||||
if (trace_flags & TRACE_ITER_LATENCY_FMT) {
|
||||
/* print nothing if the buffers are empty */
|
||||
if (trace_empty(iter))
|
||||
return;
|
||||
|
||||
print_trace_header(s, iter);
|
||||
flags |= TRACE_GRAPH_PRINT_DURATION;
|
||||
} else
|
||||
flags |= TRACE_GRAPH_PRINT_ABS_TIME;
|
||||
|
||||
print_graph_headers_flags(s, flags);
|
||||
} else
|
||||
if (is_graph())
|
||||
print_graph_headers_flags(s, GRAPH_TRACER_FLAGS);
|
||||
else
|
||||
trace_default_header(s);
|
||||
}
|
||||
|
||||
static void
|
||||
trace_graph_function(struct trace_array *tr,
|
||||
unsigned long ip, unsigned long flags, int pc)
|
||||
{
|
||||
u64 time = trace_clock_local();
|
||||
struct ftrace_graph_ent ent = {
|
||||
.func = ip,
|
||||
.depth = 0,
|
||||
};
|
||||
struct ftrace_graph_ret ret = {
|
||||
.func = ip,
|
||||
.depth = 0,
|
||||
.calltime = time,
|
||||
.rettime = time,
|
||||
};
|
||||
|
||||
__trace_graph_entry(tr, &ent, flags, pc);
|
||||
__trace_graph_return(tr, &ret, flags, pc);
|
||||
}
|
||||
|
||||
static void
|
||||
__trace_function(struct trace_array *tr,
|
||||
unsigned long ip, unsigned long parent_ip,
|
||||
unsigned long flags, int pc)
|
||||
{
|
||||
if (!is_graph())
|
||||
if (is_graph())
|
||||
trace_graph_function(tr, ip, parent_ip, flags, pc);
|
||||
else
|
||||
trace_function(tr, ip, parent_ip, flags, pc);
|
||||
else {
|
||||
trace_graph_function(tr, parent_ip, flags, pc);
|
||||
trace_graph_function(tr, ip, flags, pc);
|
||||
}
|
||||
}
|
||||
|
||||
#else
|
||||
|
@@ -31,13 +31,81 @@ static int wakeup_rt;
|
||||
static arch_spinlock_t wakeup_lock =
|
||||
(arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
|
||||
|
||||
static void wakeup_reset(struct trace_array *tr);
|
||||
static void __wakeup_reset(struct trace_array *tr);
|
||||
static int wakeup_graph_entry(struct ftrace_graph_ent *trace);
|
||||
static void wakeup_graph_return(struct ftrace_graph_ret *trace);
|
||||
|
||||
static int save_lat_flag;
|
||||
|
||||
#define TRACE_DISPLAY_GRAPH 1
|
||||
|
||||
static struct tracer_opt trace_opts[] = {
|
||||
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||
/* display latency trace as call graph */
|
||||
{ TRACER_OPT(display-graph, TRACE_DISPLAY_GRAPH) },
|
||||
#endif
|
||||
{ } /* Empty entry */
|
||||
};
|
||||
|
||||
static struct tracer_flags tracer_flags = {
|
||||
.val = 0,
|
||||
.opts = trace_opts,
|
||||
};
|
||||
|
||||
#define is_graph() (tracer_flags.val & TRACE_DISPLAY_GRAPH)
|
||||
|
||||
#ifdef CONFIG_FUNCTION_TRACER
|
||||
|
||||
/*
|
||||
* irqsoff uses its own tracer function to keep the overhead down:
|
||||
* Prologue for the wakeup function tracers.
|
||||
*
|
||||
* Returns 1 if it is OK to continue, and preemption
|
||||
* is disabled and data->disabled is incremented.
|
||||
* 0 if the trace is to be ignored, and preemption
|
||||
* is not disabled and data->disabled is
|
||||
* kept the same.
|
||||
*
|
||||
* Note, this function is also used outside this ifdef but
|
||||
* inside the #ifdef of the function graph tracer below.
|
||||
* This is OK, since the function graph tracer is
|
||||
* dependent on the function tracer.
|
||||
*/
|
||||
static int
|
||||
func_prolog_preempt_disable(struct trace_array *tr,
|
||||
struct trace_array_cpu **data,
|
||||
int *pc)
|
||||
{
|
||||
long disabled;
|
||||
int cpu;
|
||||
|
||||
if (likely(!wakeup_task))
|
||||
return 0;
|
||||
|
||||
*pc = preempt_count();
|
||||
preempt_disable_notrace();
|
||||
|
||||
cpu = raw_smp_processor_id();
|
||||
if (cpu != wakeup_current_cpu)
|
||||
goto out_enable;
|
||||
|
||||
*data = tr->data[cpu];
|
||||
disabled = atomic_inc_return(&(*data)->disabled);
|
||||
if (unlikely(disabled != 1))
|
||||
goto out;
|
||||
|
||||
return 1;
|
||||
|
||||
out:
|
||||
atomic_dec(&(*data)->disabled);
|
||||
|
||||
out_enable:
|
||||
preempt_enable_notrace();
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* wakeup uses its own tracer function to keep the overhead down:
|
||||
*/
|
||||
static void
|
||||
wakeup_tracer_call(unsigned long ip, unsigned long parent_ip)
|
||||
@@ -45,34 +113,16 @@ wakeup_tracer_call(unsigned long ip, unsigned long parent_ip)
|
||||
struct trace_array *tr = wakeup_trace;
|
||||
struct trace_array_cpu *data;
|
||||
unsigned long flags;
|
||||
long disabled;
|
||||
int cpu;
|
||||
int pc;
|
||||
|
||||
if (likely(!wakeup_task))
|
||||
if (!func_prolog_preempt_disable(tr, &data, &pc))
|
||||
return;
|
||||
|
||||
pc = preempt_count();
|
||||
preempt_disable_notrace();
|
||||
|
||||
cpu = raw_smp_processor_id();
|
||||
if (cpu != wakeup_current_cpu)
|
||||
goto out_enable;
|
||||
|
||||
data = tr->data[cpu];
|
||||
disabled = atomic_inc_return(&data->disabled);
|
||||
if (unlikely(disabled != 1))
|
||||
goto out;
|
||||
|
||||
local_irq_save(flags);
|
||||
|
||||
trace_function(tr, ip, parent_ip, flags, pc);
|
||||
|
||||
local_irq_restore(flags);
|
||||
|
||||
out:
|
||||
atomic_dec(&data->disabled);
|
||||
out_enable:
|
||||
preempt_enable_notrace();
|
||||
}
|
||||
|
||||
@@ -82,6 +132,156 @@ static struct ftrace_ops trace_ops __read_mostly =
|
||||
};
|
||||
#endif /* CONFIG_FUNCTION_TRACER */
|
||||
|
||||
static int start_func_tracer(int graph)
|
||||
{
|
||||
int ret;
|
||||
|
||||
if (!graph)
|
||||
ret = register_ftrace_function(&trace_ops);
|
||||
else
|
||||
ret = register_ftrace_graph(&wakeup_graph_return,
|
||||
&wakeup_graph_entry);
|
||||
|
||||
if (!ret && tracing_is_enabled())
|
||||
tracer_enabled = 1;
|
||||
else
|
||||
tracer_enabled = 0;
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void stop_func_tracer(int graph)
|
||||
{
|
||||
tracer_enabled = 0;
|
||||
|
||||
if (!graph)
|
||||
unregister_ftrace_function(&trace_ops);
|
||||
else
|
||||
unregister_ftrace_graph();
|
||||
}
|
||||
|
||||
#ifdef CONFIG_FUNCTION_GRAPH_TRACER
|
||||
static int wakeup_set_flag(u32 old_flags, u32 bit, int set)
|
||||
{
|
||||
|
||||
if (!(bit & TRACE_DISPLAY_GRAPH))
|
||||
return -EINVAL;
|
||||
|
||||
if (!(is_graph() ^ set))
|
||||
return 0;
|
||||
|
||||
stop_func_tracer(!set);
|
||||
|
||||
wakeup_reset(wakeup_trace);
|
||||
tracing_max_latency = 0;
|
||||
|
||||
return start_func_tracer(set);
|
||||
}
|
||||
|
||||
static int wakeup_graph_entry(struct ftrace_graph_ent *trace)
|
||||
{
|
||||
struct trace_array *tr = wakeup_trace;
|
||||
struct trace_array_cpu *data;
|
||||
unsigned long flags;
|
||||
int pc, ret = 0;
|
||||
|
||||
if (!func_prolog_preempt_disable(tr, &data, &pc))
|
||||
return 0;
|
||||
|
||||
local_save_flags(flags);
|
||||
ret = __trace_graph_entry(tr, trace, flags, pc);
|
||||
atomic_dec(&data->disabled);
|
||||
preempt_enable_notrace();
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static void wakeup_graph_return(struct ftrace_graph_ret *trace)
|
||||
{
|
||||
struct trace_array *tr = wakeup_trace;
|
||||
struct trace_array_cpu *data;
|
||||
unsigned long flags;
|
||||
int pc;
|
||||
|
||||
if (!func_prolog_preempt_disable(tr, &data, &pc))
|
||||
return;
|
||||
|
||||
local_save_flags(flags);
|
||||
__trace_graph_return(tr, trace, flags, pc);
|
||||
atomic_dec(&data->disabled);
|
||||
|
||||
preempt_enable_notrace();
|
||||
return;
|
||||
}
|
||||
|
||||
static void wakeup_trace_open(struct trace_iterator *iter)
|
||||
{
|
||||
if (is_graph())
|
||||
graph_trace_open(iter);
|
||||
}
|
||||
|
||||
static void wakeup_trace_close(struct trace_iterator *iter)
|
||||
{
|
||||
if (iter->private)
|
||||
graph_trace_close(iter);
|
||||
}
|
||||
|
||||
#define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_PROC)
|
||||
|
||||
static enum print_line_t wakeup_print_line(struct trace_iterator *iter)
|
||||
{
|
||||
/*
|
||||
* In graph mode call the graph tracer output function,
|
||||
* otherwise go with the TRACE_FN event handler
|
||||
*/
|
||||
if (is_graph())
|
||||
return print_graph_function_flags(iter, GRAPH_TRACER_FLAGS);
|
||||
|
||||
return TRACE_TYPE_UNHANDLED;
|
||||
}
|
||||
|
||||
static void wakeup_print_header(struct seq_file *s)
|
||||
{
|
||||
if (is_graph())
|
||||
print_graph_headers_flags(s, GRAPH_TRACER_FLAGS);
|
||||
else
|
||||
trace_default_header(s);
|
||||
}
|
||||
|
||||
static void
|
||||
__trace_function(struct trace_array *tr,
|
||||
unsigned long ip, unsigned long parent_ip,
|
||||
unsigned long flags, int pc)
|
||||
{
|
||||
if (is_graph())
|
||||
trace_graph_function(tr, ip, parent_ip, flags, pc);
|
||||
else
|
||||
trace_function(tr, ip, parent_ip, flags, pc);
|
||||
}
|
||||
#else
|
||||
#define __trace_function trace_function
|
||||
|
||||
static int wakeup_set_flag(u32 old_flags, u32 bit, int set)
|
||||
{
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
static int wakeup_graph_entry(struct ftrace_graph_ent *trace)
|
||||
{
|
||||
return -1;
|
||||
}
|
||||
|
||||
static enum print_line_t wakeup_print_line(struct trace_iterator *iter)
|
||||
{
|
||||
return TRACE_TYPE_UNHANDLED;
|
||||
}
|
||||
|
||||
static void wakeup_graph_return(struct ftrace_graph_ret *trace) { }
|
||||
static void wakeup_print_header(struct seq_file *s) { }
|
||||
static void wakeup_trace_open(struct trace_iterator *iter) { }
|
||||
static void wakeup_trace_close(struct trace_iterator *iter) { }
|
||||
#endif /* CONFIG_FUNCTION_GRAPH_TRACER */
|
||||
|
||||
/*
|
||||
* Should this new latency be reported/recorded?
|
||||
*/
|
||||
@@ -152,7 +352,7 @@ probe_wakeup_sched_switch(void *ignore,
|
||||
/* The task we are waiting for is waking up */
|
||||
data = wakeup_trace->data[wakeup_cpu];
|
||||
|
||||
trace_function(wakeup_trace, CALLER_ADDR0, CALLER_ADDR1, flags, pc);
|
||||
__trace_function(wakeup_trace, CALLER_ADDR0, CALLER_ADDR1, flags, pc);
|
||||
tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc);
|
||||
|
||||
T0 = data->preempt_timestamp;
|
||||
@@ -252,7 +452,7 @@ probe_wakeup(void *ignore, struct task_struct *p, int success)
|
||||
* is not called by an assembly function (where as schedule is)
|
||||
* it should be safe to use it here.
|
||||
*/
|
||||
trace_function(wakeup_trace, CALLER_ADDR1, CALLER_ADDR2, flags, pc);
|
||||
__trace_function(wakeup_trace, CALLER_ADDR1, CALLER_ADDR2, flags, pc);
|
||||
|
||||
out_locked:
|
||||
arch_spin_unlock(&wakeup_lock);
|
||||
@@ -303,12 +503,8 @@ static void start_wakeup_tracer(struct trace_array *tr)
|
||||
*/
|
||||
smp_wmb();
|
||||
|
||||
register_ftrace_function(&trace_ops);
|
||||
|
||||
if (tracing_is_enabled())
|
||||
tracer_enabled = 1;
|
||||
else
|
||||
tracer_enabled = 0;
|
||||
if (start_func_tracer(is_graph()))
|
||||
printk(KERN_ERR "failed to start wakeup tracer\n");
|
||||
|
||||
return;
|
||||
fail_deprobe_wake_new:
|
||||
@@ -320,7 +516,7 @@ fail_deprobe:
|
||||
static void stop_wakeup_tracer(struct trace_array *tr)
|
||||
{
|
||||
tracer_enabled = 0;
|
||||
unregister_ftrace_function(&trace_ops);
|
||||
stop_func_tracer(is_graph());
|
||||
unregister_trace_sched_switch(probe_wakeup_sched_switch, NULL);
|
||||
unregister_trace_sched_wakeup_new(probe_wakeup, NULL);
|
||||
unregister_trace_sched_wakeup(probe_wakeup, NULL);
|
||||
@@ -379,9 +575,15 @@ static struct tracer wakeup_tracer __read_mostly =
|
||||
.start = wakeup_tracer_start,
|
||||
.stop = wakeup_tracer_stop,
|
||||
.print_max = 1,
|
||||
.print_header = wakeup_print_header,
|
||||
.print_line = wakeup_print_line,
|
||||
.flags = &tracer_flags,
|
||||
.set_flag = wakeup_set_flag,
|
||||
#ifdef CONFIG_FTRACE_SELFTEST
|
||||
.selftest = trace_selftest_startup_wakeup,
|
||||
#endif
|
||||
.open = wakeup_trace_open,
|
||||
.close = wakeup_trace_close,
|
||||
.use_max_tr = 1,
|
||||
};
|
||||
|
||||
@@ -394,9 +596,15 @@ static struct tracer wakeup_rt_tracer __read_mostly =
|
||||
.stop = wakeup_tracer_stop,
|
||||
.wait_pipe = poll_wait_pipe,
|
||||
.print_max = 1,
|
||||
.print_header = wakeup_print_header,
|
||||
.print_line = wakeup_print_line,
|
||||
.flags = &tracer_flags,
|
||||
.set_flag = wakeup_set_flag,
|
||||
#ifdef CONFIG_FTRACE_SELFTEST
|
||||
.selftest = trace_selftest_startup_wakeup,
|
||||
#endif
|
||||
.open = wakeup_trace_open,
|
||||
.close = wakeup_trace_close,
|
||||
.use_max_tr = 1,
|
||||
};
|
||||
|
||||
|
@@ -263,6 +263,11 @@ int __init trace_workqueue_early_init(void)
|
||||
{
|
||||
int ret, cpu;
|
||||
|
||||
for_each_possible_cpu(cpu) {
|
||||
spin_lock_init(&workqueue_cpu_stat(cpu)->lock);
|
||||
INIT_LIST_HEAD(&workqueue_cpu_stat(cpu)->list);
|
||||
}
|
||||
|
||||
ret = register_trace_workqueue_insertion(probe_workqueue_insertion, NULL);
|
||||
if (ret)
|
||||
goto out;
|
||||
@@ -279,11 +284,6 @@ int __init trace_workqueue_early_init(void)
|
||||
if (ret)
|
||||
goto no_creation;
|
||||
|
||||
for_each_possible_cpu(cpu) {
|
||||
spin_lock_init(&workqueue_cpu_stat(cpu)->lock);
|
||||
INIT_LIST_HEAD(&workqueue_cpu_stat(cpu)->list);
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
||||
no_creation:
|
||||
|
Reference in New Issue
Block a user