Skip to content

Commit

Permalink
---
Browse files Browse the repository at this point in the history
yaml
---
r: 121053
b: refs/heads/master
c: 182e9f5
h: refs/heads/master
i:
  121051: 1d74e55
v: v3
  • Loading branch information
Steven Rostedt authored and Ingo Molnar committed Nov 4, 2008
1 parent be664c8 commit 6bfbf66
Show file tree
Hide file tree
Showing 5 changed files with 16 additions and 42 deletions.
2 changes: 1 addition & 1 deletion [refs]
Original file line number Diff line number Diff line change
@@ -1,2 +1,2 @@
---
refs/heads/master: 8f0a056fcb2f83a069fb5d60c2383304b7456687
refs/heads/master: 182e9f5f704ed6b9175142fe8da33c9ce0c52b52
27 changes: 9 additions & 18 deletions trunk/kernel/trace/ring_buffer.c
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,8 @@
#include <linux/list.h>
#include <linux/fs.h>

#include "trace.h"

/* Up this if you want to test the TIME_EXTENTS and normalization */
#define DEBUG_SHIFT 0

Expand Down Expand Up @@ -1122,8 +1124,7 @@ ring_buffer_lock_reserve(struct ring_buffer *buffer,
return NULL;

/* If we are tracing schedule, we don't want to recurse */
resched = need_resched();
preempt_disable_notrace();
resched = ftrace_preempt_disable();

cpu = raw_smp_processor_id();

Expand Down Expand Up @@ -1154,10 +1155,7 @@ ring_buffer_lock_reserve(struct ring_buffer *buffer,
return event;

out:
if (resched)
preempt_enable_notrace();
else
preempt_enable_notrace();
ftrace_preempt_enable(resched);
return NULL;
}

Expand Down Expand Up @@ -1199,12 +1197,9 @@ int ring_buffer_unlock_commit(struct ring_buffer *buffer,
/*
* Only the last preempt count needs to restore preemption.
*/
if (preempt_count() == 1) {
if (per_cpu(rb_need_resched, cpu))
preempt_enable_no_resched_notrace();
else
preempt_enable_notrace();
} else
if (preempt_count() == 1)
ftrace_preempt_enable(per_cpu(rb_need_resched, cpu));
else
preempt_enable_no_resched_notrace();

return 0;
Expand Down Expand Up @@ -1237,8 +1232,7 @@ int ring_buffer_write(struct ring_buffer *buffer,
if (atomic_read(&buffer->record_disabled))
return -EBUSY;

resched = need_resched();
preempt_disable_notrace();
resched = ftrace_preempt_disable();

cpu = raw_smp_processor_id();

Expand All @@ -1264,10 +1258,7 @@ int ring_buffer_write(struct ring_buffer *buffer,

ret = 0;
out:
if (resched)
preempt_enable_no_resched_notrace();
else
preempt_enable_notrace();
ftrace_preempt_enable(resched);

return ret;
}
Expand Down
8 changes: 2 additions & 6 deletions trunk/kernel/trace/trace.c
Original file line number Diff line number Diff line change
Expand Up @@ -904,8 +904,7 @@ function_trace_call(unsigned long ip, unsigned long parent_ip)
return;

pc = preempt_count();
resched = need_resched();
preempt_disable_notrace();
resched = ftrace_preempt_disable();
local_save_flags(flags);
cpu = raw_smp_processor_id();
data = tr->data[cpu];
Expand All @@ -915,10 +914,7 @@ function_trace_call(unsigned long ip, unsigned long parent_ip)
trace_function(tr, data, ip, parent_ip, flags, pc);

atomic_dec(&data->disabled);
if (resched)
preempt_enable_no_resched_notrace();
else
preempt_enable_notrace();
ftrace_preempt_enable(resched);
}

static struct ftrace_ops trace_ops __read_mostly =
Expand Down
13 changes: 2 additions & 11 deletions trunk/kernel/trace/trace_sched_wakeup.c
Original file line number Diff line number Diff line change
Expand Up @@ -50,8 +50,7 @@ wakeup_tracer_call(unsigned long ip, unsigned long parent_ip)
return;

pc = preempt_count();
resched = need_resched();
preempt_disable_notrace();
resched = ftrace_preempt_disable();

cpu = raw_smp_processor_id();
data = tr->data[cpu];
Expand Down Expand Up @@ -81,15 +80,7 @@ wakeup_tracer_call(unsigned long ip, unsigned long parent_ip)
out:
atomic_dec(&data->disabled);

/*
* To prevent recursion from the scheduler, if the
* resched flag was set before we entered, then
* don't reschedule.
*/
if (resched)
preempt_enable_no_resched_notrace();
else
preempt_enable_notrace();
ftrace_preempt_enable(resched);
}

static struct ftrace_ops trace_ops __read_mostly =
Expand Down
8 changes: 2 additions & 6 deletions trunk/kernel/trace/trace_stack.c
Original file line number Diff line number Diff line change
Expand Up @@ -107,8 +107,7 @@ stack_trace_call(unsigned long ip, unsigned long parent_ip)
if (unlikely(!ftrace_enabled || stack_trace_disabled))
return;

resched = need_resched();
preempt_disable_notrace();
resched = ftrace_preempt_disable();

cpu = raw_smp_processor_id();
/* no atomic needed, we only modify this variable by this cpu */
Expand All @@ -120,10 +119,7 @@ stack_trace_call(unsigned long ip, unsigned long parent_ip)
out:
per_cpu(trace_active, cpu)--;
/* prevent recursion in schedule */
if (resched)
preempt_enable_no_resched_notrace();
else
preempt_enable_notrace();
ftrace_preempt_enable(resched);
}

static struct ftrace_ops trace_ops __read_mostly =
Expand Down

0 comments on commit 6bfbf66

Please sign in to comment.