The freeing_list cleanup now retries optimizing any sibling probe that was
deferred while this aggregator was being torn down.  Track the pending
address in struct optimized_kprobe so __disarm_kprobe() can defer the
retry until kprobe_optimizer() finishes disarming.

Signed-off-by: hongao <[email protected]>
---
Changes since v1:
- Replace `kprobe_opcode_t *pending_reopt_addr` with `bool 
reopt_unblocked_probes`
  in `struct optimized_kprobe` to avoid storing an address and simplify logic.
- Use `op->kp.addr` when looking up the sibling optimized probe instead of
  keeping a separate stored address.
- Defer re-optimization by setting/clearing `op->reopt_unblocked_probes` in
  `__disarm_kprobe()` / consuming it in `do_free_cleaned_kprobes()` so the
  retry runs after the worker finishes disarming.
- Link to v1: 
https://lore.kernel.org/all/[email protected]/
---
 include/linux/kprobes.h |  1 +
 kernel/kprobes.c        | 28 ++++++++++++++++++++++------
 2 files changed, 23 insertions(+), 6 deletions(-)

diff --git a/include/linux/kprobes.h b/include/linux/kprobes.h
index 8c4f3bb24..4f49925a4 100644
--- a/include/linux/kprobes.h
+++ b/include/linux/kprobes.h
@@ -338,6 +338,7 @@ DEFINE_INSN_CACHE_OPS(insn);
 struct optimized_kprobe {
        struct kprobe kp;
        struct list_head list;  /* list for optimizing queue */
+       bool reopt_unblocked_probes;
        struct arch_optimized_insn optinsn;
 };
 
diff --git a/kernel/kprobes.c b/kernel/kprobes.c
index da59c68df..799542dff 100644
--- a/kernel/kprobes.c
+++ b/kernel/kprobes.c
@@ -514,6 +514,7 @@ static LIST_HEAD(freeing_list);
 
 static void kprobe_optimizer(struct work_struct *work);
 static DECLARE_DELAYED_WORK(optimizing_work, kprobe_optimizer);
+static void optimize_kprobe(struct kprobe *p);
 #define OPTIMIZE_DELAY 5
 
 /*
@@ -591,6 +592,21 @@ static void do_free_cleaned_kprobes(void)
                         */
                        continue;
                }
+               if (op->reopt_unblocked_probes) {
+                       struct kprobe *unblocked;
+
+                       /*
+                        * The aggregator was holding back another probe while 
it sat on the
+                        * unoptimizing/freeing lists.  Now that the aggregator 
has been fully
+                        * reverted we can safely retry the optimization of 
that sibling.
+                        */
+
+                       unblocked = get_optimized_kprobe(op->kp.addr);
+                       if (unlikely(unblocked))
+                               optimize_kprobe(unblocked);
+                       op->reopt_unblocked_probes = false;
+               }
+
                free_aggr_kprobe(&op->kp);
        }
 }
@@ -1009,13 +1025,13 @@ static void __disarm_kprobe(struct kprobe *p, bool 
reopt)
                _p = get_optimized_kprobe(p->addr);
                if (unlikely(_p) && reopt)
                        optimize_kprobe(_p);
+       } else if (reopt && kprobe_aggrprobe(p)) {
+               struct optimized_kprobe *op =
+                       container_of(p, struct optimized_kprobe, kp);
+
+               /* Defer the re-optimization until the worker finishes 
disarming. */
+               op->reopt_unblocked_probes = true;
        }
-       /*
-        * TODO: Since unoptimization and real disarming will be done by
-        * the worker thread, we can not check whether another probe are
-        * unoptimized because of this probe here. It should be re-optimized
-        * by the worker thread.
-        */
 }
 
 #else /* !CONFIG_OPTPROBES */
-- 
2.47.2


Reply via email to