Hopefully this is the final version before I push this to linux-next.

Paul, can I get an ack on the last patch from you?

-- Steve

Paul E. McKenney (1):
      rcu: Fix dyntick-idle tracing

Steven Rostedt (VMware) (6):
      ftrace: Add use of synchronize_rcu_tasks() with dynamic trampolines
      tracing: Replace the per_cpu() with __this_cpu*() in trace_stack.c
      tracing: Add stack_tracer_disable/enable() functions
      tracing: Rename trace_active to disable_stack_tracer and inline its 
modification
      rcu/tracing: Add rcu_disabled to denote when rcu_irq_enter() will not work
      tracing: Make sure rcu_irq_enter() can work for trace_*_rcuidle() trace 
events

----
 include/linux/ftrace.h     | 38 ++++++++++++++++++++++++++++
 include/linux/rcupdate.h   |  5 ++++
 include/linux/tracepoint.h | 19 ++++++++------
 kernel/rcu/tree.c          | 62 +++++++++++++++++++++++++++-------------------
 kernel/trace/Kconfig       |  3 ++-
 kernel/trace/ftrace.c      | 42 ++++++++++++++-----------------
 kernel/trace/trace_stack.c | 35 +++++++++++++-------------
 7 files changed, 128 insertions(+), 76 deletions(-)

Diff against v2 (not counting the last two patches at the end of this series)

diff --git a/include/linux/ftrace.h b/include/linux/ftrace.h
index 4bde7ff..06b2990 100644
--- a/include/linux/ftrace.h
+++ b/include/linux/ftrace.h
@@ -323,7 +323,7 @@ static inline void stack_tracer_enable(void)
 }
 #else
 static inline void stack_tracer_disable(void) { }
-static inline void stack_tracer_enabe(void) { }
+static inline void stack_tracer_enable(void) { }
 #endif
 
 struct ftrace_func_command {
diff --git a/kernel/trace/trace_stack.c b/kernel/trace/trace_stack.c
index 84fafb6..f2f02ff 100644
--- a/kernel/trace/trace_stack.c
+++ b/kernel/trace/trace_stack.c
@@ -211,8 +211,8 @@ stack_trace_call(unsigned long ip, unsigned long parent_ip,
        preempt_disable_notrace();
 
        /* no atomic needed, we only modify this variable by this cpu */
-       this_cpu_inc(disable_stack_tracer);
-       if (this_cpu_read(disable_stack_tracer) != 1)
+       __this_cpu_inc(disable_stack_tracer);
+       if (__this_cpu_read(disable_stack_tracer) != 1)
                goto out;
 
        ip += MCOUNT_INSN_SIZE;
@@ -220,7 +220,7 @@ stack_trace_call(unsigned long ip, unsigned long parent_ip,
        check_stack(ip, &stack);
 
  out:
-       this_cpu_dec(disable_stack_tracer);
+       __this_cpu_dec(disable_stack_tracer);
        /* prevent recursion in schedule */
        preempt_enable_notrace();
 }
@@ -264,13 +264,13 @@ stack_max_size_write(struct file *filp, const char __user 
*ubuf,
         * we will cause circular lock, so we also need to increase
         * the percpu disable_stack_tracer here.
         */
-       this_cpu_inc(disable_stack_tracer);
+       __this_cpu_inc(disable_stack_tracer);
 
        arch_spin_lock(&stack_trace_max_lock);
        *ptr = val;
        arch_spin_unlock(&stack_trace_max_lock);
 
-       this_cpu_dec(disable_stack_tracer);
+       __this_cpu_dec(disable_stack_tracer);
        local_irq_restore(flags);
 
        return count;
@@ -306,7 +306,7 @@ static void *t_start(struct seq_file *m, loff_t *pos)
 {
        local_irq_disable();
 
-       this_cpu_inc(disable_stack_tracer);
+       __this_cpu_inc(disable_stack_tracer);
 
        arch_spin_lock(&stack_trace_max_lock);
 
@@ -320,7 +320,7 @@ static void t_stop(struct seq_file *m, void *p)
 {
        arch_spin_unlock(&stack_trace_max_lock);
 
-       this_cpu_dec(disable_stack_tracer);
+       __this_cpu_dec(disable_stack_tracer);
 
        local_irq_enable();
 }

Reply via email to