> 
> I think it would be sufficient to call RBCOMPUTE(node, true) on every
> node and check the return value ?
>
Yes, that is enough for sure. The only way i was thinking about to make it
public, because checking the tree for MAX is generic for every users which
use RB_DECLARE_CALLBACKS_MAX template. Something like:

validate_rb_max_tree() {
    for (nd = rb_first(root); nd; nd = rb_next(nd)) {
            fooo = rb_entry(nd, struct sometinhf, rb_field);
            WARN_ON(!*_compute_max(foo, true);  
    }
}

and call this public function under debug code. But i do not have strong
opinion here and it is probably odd. Anyway i am fine with your change.

There is small comment below:

> 
> Something like the following (probably applicable in other files too):
> 
> ---------------------------------- 8< ------------------------------------
> 
> augmented rbtree: use generated compute_max function for debug checks
> 
> In debug code, use the generated compute_max function instead of
> reimplementing similar functionality in multiple places.
> 
> Signed-off-by: Michel Lespinasse <wal...@google.com>
> ---
>  lib/rbtree_test.c | 15 +-------------
>  mm/mmap.c         | 26 +++--------------------
>  mm/vmalloc.c      | 53 +++++++----------------------------------------
>  3 files changed, 12 insertions(+), 82 deletions(-)
> 
> diff --git a/lib/rbtree_test.c b/lib/rbtree_test.c
> index 41ae3c7570d3..a5a04e820f77 100644
> --- a/lib/rbtree_test.c
> +++ b/lib/rbtree_test.c
> @@ -222,20 +222,7 @@ static void check_augmented(int nr_nodes)
>       check(nr_nodes);
>       for (rb = rb_first(&root.rb_root); rb; rb = rb_next(rb)) {
>               struct test_node *node = rb_entry(rb, struct test_node, rb);
> -             u32 subtree, max = node->val;
> -             if (node->rb.rb_left) {
> -                     subtree = rb_entry(node->rb.rb_left, struct test_node,
> -                                        rb)->augmented;
> -                     if (max < subtree)
> -                             max = subtree;
> -             }
> -             if (node->rb.rb_right) {
> -                     subtree = rb_entry(node->rb.rb_right, struct test_node,
> -                                        rb)->augmented;
> -                     if (max < subtree)
> -                             max = subtree;
> -             }
> -             WARN_ON_ONCE(node->augmented != max);
> +             WARN_ON_ONCE(!augment_callbacks_compute_max(node, true));
>       }
>  }
>  
> diff --git a/mm/mmap.c b/mm/mmap.c
> index 24f0772d6afd..d6d23e6c2d10 100644
> --- a/mm/mmap.c
> +++ b/mm/mmap.c
> @@ -311,24 +311,6 @@ static inline unsigned long vma_compute_gap(struct 
> vm_area_struct *vma)
>  }
>  
>  #ifdef CONFIG_DEBUG_VM_RB
> -static unsigned long vma_compute_subtree_gap(struct vm_area_struct *vma)
> -{
> -     unsigned long max = vma_compute_gap(vma), subtree_gap;
> -     if (vma->vm_rb.rb_left) {
> -             subtree_gap = rb_entry(vma->vm_rb.rb_left,
> -                             struct vm_area_struct, vm_rb)->rb_subtree_gap;
> -             if (subtree_gap > max)
> -                     max = subtree_gap;
> -     }
> -     if (vma->vm_rb.rb_right) {
> -             subtree_gap = rb_entry(vma->vm_rb.rb_right,
> -                             struct vm_area_struct, vm_rb)->rb_subtree_gap;
> -             if (subtree_gap > max)
> -                     max = subtree_gap;
> -     }
> -     return max;
> -}
> -
>  static int browse_rb(struct mm_struct *mm)
>  {
>       struct rb_root *root = &mm->mm_rb;
> @@ -355,10 +337,8 @@ static int browse_rb(struct mm_struct *mm)
>                       bug = 1;
>               }
>               spin_lock(&mm->page_table_lock);
> -             if (vma->rb_subtree_gap != vma_compute_subtree_gap(vma)) {
> -                     pr_emerg("free gap %lx, correct %lx\n",
> -                            vma->rb_subtree_gap,
> -                            vma_compute_subtree_gap(vma));
> +             if (!vma_gap_callbacks_compute_max(vma, true)) {
> +                     pr_emerg("wrong subtree gap in vma %p\n", vma);
>                       bug = 1;
>               }
>               spin_unlock(&mm->page_table_lock);
> @@ -385,7 +365,7 @@ static void validate_mm_rb(struct rb_root *root, struct 
> vm_area_struct *ignore)
>               struct vm_area_struct *vma;
>               vma = rb_entry(nd, struct vm_area_struct, vm_rb);
>               VM_BUG_ON_VMA(vma != ignore &&
> -                     vma->rb_subtree_gap != vma_compute_subtree_gap(vma),
> +                     !vma_gap_callbacks_compute_max(vma, true),
>                       vma);
>       }
>  }
> diff --git a/mm/vmalloc.c b/mm/vmalloc.c
> index f7c61accb0e2..ea23ccaf70fc 100644
> --- a/mm/vmalloc.c
> +++ b/mm/vmalloc.c
> @@ -553,48 +553,6 @@ unlink_va(struct vmap_area *va, struct rb_root *root)
>       RB_CLEAR_NODE(&va->rb_node);
>  }
>  
> -#if DEBUG_AUGMENT_PROPAGATE_CHECK
> -static void
> -augment_tree_propagate_check(struct rb_node *n)
> -{
> -     struct vmap_area *va;
> -     struct rb_node *node;
> -     unsigned long size;
> -     bool found = false;
> -
> -     if (n == NULL)
> -             return;
> -
> -     va = rb_entry(n, struct vmap_area, rb_node);
> -     size = va->subtree_max_size;
> -     node = n;
> -
> -     while (node) {
> -             va = rb_entry(node, struct vmap_area, rb_node);
> -
> -             if (get_subtree_max_size(node->rb_left) == size) {
> -                     node = node->rb_left;
> -             } else {
> -                     if (va_size(va) == size) {
> -                             found = true;
> -                             break;
> -                     }
> -
> -                     node = node->rb_right;
> -             }
> -     }
> -
> -     if (!found) {
> -             va = rb_entry(n, struct vmap_area, rb_node);
> -             pr_emerg("tree is corrupted: %lu, %lu\n",
> -                     va_size(va), va->subtree_max_size);
> -     }
> -
> -     augment_tree_propagate_check(n->rb_left);
> -     augment_tree_propagate_check(n->rb_right);
> -}
> -#endif
> -
>  /*
>   * This function populates subtree_max_size from bottom to upper
>   * levels starting from VA point. The propagation must be done
> @@ -645,9 +603,14 @@ augment_tree_propagate_from(struct vmap_area *va)
>               node = rb_parent(&va->rb_node);
>       }
>  
> -#if DEBUG_AUGMENT_PROPAGATE_CHECK
> -     augment_tree_propagate_check(free_vmap_area_root.rb_node);
> -#endif
> +     if (DEBUG_AUGMENT_PROPAGATE_CHECK) {
> +             struct vmap_area *va;
> +
> +             list_for_each_entry(va, &free_vmap_area_list, list) {
> +                     WARN_ON(!free_vmap_area_rb_augment_cb_compute_max(
> +                                     va, true));
> +             }
> +     }
>  }
>
The object of validating is the tree, therefore it makes sense to go with it,
instead of iterating over the list.

Thank you!

--
Vlad Rezki

Reply via email to