On 08/09, Steven Rostedt wrote:
>
> --- a/kernel/trace/trace_uprobe.c
> +++ b/kernel/trace/trace_uprobe.c
> @@ -952,7 +952,7 @@ probe_event_disable(struct trace_uprobe *tu, struct 
> trace_event_file *file)
>
>               list_del_rcu(&link->list);
>               /* synchronize with u{,ret}probe_trace_func */
> -             synchronize_sched();
> +             synchronize_rcu();

Can't we change uprobe_trace_func() and uretprobe_trace_func() to use
rcu_read_lock_sched() instead? It is more cheap.


Hmm. probe_event_enable() does list_del + kfree on failure, this doesn't
look right... Not only because kfree() can race with list_for_each_entry_rcu(),
we should not put the 1st link on list until uprobe_buffer_enable().

Does the patch below make sense or I am confused?

Oleg.


--- x/kernel/trace/trace_uprobe.c
+++ x/kernel/trace/trace_uprobe.c
@@ -896,8 +896,6 @@ probe_event_enable(struct trace_uprobe *
                        return -ENOMEM;
 
                link->file = file;
-               list_add_tail_rcu(&link->list, &tu->tp.files);
-
                tu->tp.flags |= TP_FLAG_TRACE;
        } else {
                if (tu->tp.flags & TP_FLAG_TRACE)
@@ -909,7 +907,7 @@ probe_event_enable(struct trace_uprobe *
        WARN_ON(!uprobe_filter_is_empty(&tu->filter));
 
        if (enabled)
-               return 0;
+               goto add;
 
        ret = uprobe_buffer_enable();
        if (ret)
@@ -920,7 +918,8 @@ probe_event_enable(struct trace_uprobe *
        ret = uprobe_register(tu->inode, tu->offset, &tu->consumer);
        if (ret)
                goto err_buffer;
-
+ add:
+       list_add_tail_rcu(&link->list, &tu->tp.files);
        return 0;
 
  err_buffer:
@@ -928,7 +927,6 @@ probe_event_enable(struct trace_uprobe *
 
  err_flags:
        if (file) {
-               list_del(&link->list);
                kfree(link);
                tu->tp.flags &= ~TP_FLAG_TRACE;
        } else {

Reply via email to