From: "Steven Rostedt (Red Hat)" <rost...@goodmis.org>

Fix up a few typos in comments and convert an int into a bool in
update_traceon_count().

Link: http://lkml.kernel.org/r/546dd445.5080...@hitachi.com

Suggested-by: Masami Hiramatsu <masami.hiramatsu...@hitachi.com>
Signed-off-by: Steven Rostedt <rost...@goodmis.org>
---
 kernel/trace/ftrace.c          | 2 +-
 kernel/trace/trace_functions.c | 6 +++---
 2 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c
index fa0f36bb32e9..588af40d33db 100644
--- a/kernel/trace/ftrace.c
+++ b/kernel/trace/ftrace.c
@@ -1119,7 +1119,7 @@ static struct ftrace_ops global_ops = {
 
 /*
  * This is used by __kernel_text_address() to return true if the
- * the address is on a dynamically allocated trampoline that would
+ * address is on a dynamically allocated trampoline that would
  * not return true for either core_kernel_text() or
  * is_module_text_address().
  */
diff --git a/kernel/trace/trace_functions.c b/kernel/trace/trace_functions.c
index 973db52eb070..fcd41a166405 100644
--- a/kernel/trace/trace_functions.c
+++ b/kernel/trace/trace_functions.c
@@ -261,14 +261,14 @@ static struct tracer function_trace __tracer_data =
 };
 
 #ifdef CONFIG_DYNAMIC_FTRACE
-static void update_traceon_count(void **data, int on)
+static void update_traceon_count(void **data, bool on)
 {
        long *count = (long *)data;
        long old_count = *count;
 
        /*
         * Tracing gets disabled (or enabled) once per count.
-        * This function can be called at the same time on mulitple CPUs.
+        * This function can be called at the same time on multiple CPUs.
         * It is fine if both disable (or enable) tracing, as disabling
         * (or enabling) the second time doesn't do anything as the
         * state of the tracer is already disabled (or enabled).
@@ -288,7 +288,7 @@ static void update_traceon_count(void **data, int on)
         * the new state is visible before changing the counter by
         * one minus the old counter. This guarantees that another CPU
         * executing this code will see the new state before seeing
-        * the new counter value, and would not do anthing if the new
+        * the new counter value, and would not do anything if the new
         * counter is seen.
         *
         * Note, there is no synchronization between this and a user
-- 
2.1.1


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to