summaryrefslogtreecommitdiffstats
path: root/kernel/trace
diff options
context:
space:
mode:
Diffstat (limited to 'kernel/trace')
-rw-r--r--kernel/trace/ftrace.c15
1 files changed, 11 insertions, 4 deletions
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c
index 40f64f7cd85..af5ad8949ab 100644
--- a/kernel/trace/ftrace.c
+++ b/kernel/trace/ftrace.c
@@ -267,6 +267,7 @@ ftrace_record_ip(unsigned long ip)
unsigned long key;
int resched;
int atomic;
+ int cpu;
if (!ftrace_enabled || ftrace_disabled)
return;
@@ -274,9 +275,15 @@ ftrace_record_ip(unsigned long ip)
resched = need_resched();
preempt_disable_notrace();
- /* We simply need to protect against recursion */
- __get_cpu_var(ftrace_shutdown_disable_cpu)++;
- if (__get_cpu_var(ftrace_shutdown_disable_cpu) != 1)
+ /*
+ * We simply need to protect against recursion.
+ * Use the the raw version of smp_processor_id and not
+ * __get_cpu_var which can call debug hooks that can
+ * cause a recursive crash here.
+ */
+ cpu = raw_smp_processor_id();
+ per_cpu(ftrace_shutdown_disable_cpu, cpu)++;
+ if (per_cpu(ftrace_shutdown_disable_cpu, cpu) != 1)
goto out;
if (unlikely(ftrace_record_suspend))
@@ -317,7 +324,7 @@ ftrace_record_ip(unsigned long ip)
out_unlock:
spin_unlock_irqrestore(&ftrace_shutdown_lock, flags);
out:
- __get_cpu_var(ftrace_shutdown_disable_cpu)--;
+ per_cpu(ftrace_shutdown_disable_cpu, cpu)--;
/* prevent recursion with scheduler */
if (resched)