From 9f40215240f74d5b86b6f4d0ce82dcb6f7de44e3 Mon Sep 17 00:00:00 2001 From: Shan Wei Date: Mon, 19 Nov 2012 13:21:01 +0800 Subject: [PATCH] --- yaml --- r: 350038 b: refs/heads/master c: 821465295b36136998ef294fe176fba4e09c1cd9 h: refs/heads/master v: v3 --- [refs] | 2 +- trunk/kernel/trace/trace.c | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/[refs] b/[refs] index 1fd5582a358b..89a3a5f28d17 100644 --- a/[refs] +++ b/[refs] @@ -1,2 +1,2 @@ --- -refs/heads/master: d75f717e19fe595e7efbf67de195ada8d89dfbbe +refs/heads/master: 821465295b36136998ef294fe176fba4e09c1cd9 diff --git a/trunk/kernel/trace/trace.c b/trunk/kernel/trace/trace.c index ca9b7dfed8ef..07888e15c694 100644 --- a/trunk/kernel/trace/trace.c +++ b/trunk/kernel/trace/trace.c @@ -1344,7 +1344,7 @@ static void __ftrace_trace_stack(struct ring_buffer *buffer, */ preempt_disable_notrace(); - use_stack = ++__get_cpu_var(ftrace_stack_reserve); + use_stack = __this_cpu_inc_return(ftrace_stack_reserve); /* * We don't need any atomic variables, just a barrier. * If an interrupt comes in, we don't care, because it would @@ -1398,7 +1398,7 @@ static void __ftrace_trace_stack(struct ring_buffer *buffer, out: /* Again, don't let gcc optimize things here */ barrier(); - __get_cpu_var(ftrace_stack_reserve)--; + __this_cpu_dec(ftrace_stack_reserve); preempt_enable_notrace(); }