Add basic internal add/remove/search functions for btrfs_dedup.
They are all internal use only as caller shouldn't call this low level
functions

Signed-off-by: Qu Wenruo <quwen...@cn.fujitsu.com>
---
 fs/btrfs/dedup.c | 169 ++++++++++++++++++++++++++++++++++++++++++++++++++++++-
 1 file changed, 167 insertions(+), 2 deletions(-)

diff --git a/fs/btrfs/dedup.c b/fs/btrfs/dedup.c
index 41f70f5..2bce303 100644
--- a/fs/btrfs/dedup.c
+++ b/fs/btrfs/dedup.c
@@ -78,11 +78,176 @@ fail:
        return ret;
 }
 
+static void __dedup_del_hash(struct btrfs_dedup_root *root,
+                            struct btrfs_dedup_hash *hash)
+{
+       list_del(&hash->lru_list);
+       rb_erase(&hash->hash_node, &root->hash_root);
+       rb_erase(&hash->bytenr_node, &root->bytenr_root);
+
+       WARN_ON(root->current_nr == 0);
+       root->current_nr--;
+}
+
 void btrfs_free_dedup(struct btrfs_fs_info *fs_info)
 {
-       if (!fs_info->dedup_root)
+       struct btrfs_dedup_hash *entry, *tmp;
+       struct btrfs_dedup_root *dedup_root = fs_info->dedup_root;
+
+       if (!dedup_root)
                return;
 
-       kfree(fs_info->dedup_root);
+       spin_lock(&dedup_root->lock);
+       list_for_each_entry_safe(entry, tmp, &dedup_root->lru_list, lru_list) {
+               __dedup_del_hash(dedup_root, entry);
+               kmem_cache_free(btrfs_dedup_hash_cachep, entry);
+       }
+       spin_unlock(&dedup_root->lock);
+       kfree(dedup_root);
        return;
 }
+
+static int __hash_page(struct btrfs_dedup_root *root, struct page *page,
+                      char *out)
+{
+       struct crypto_shash *tfm = root->dedup_driver;
+       struct {
+               struct shash_desc desc;
+               char ctx[crypto_shash_descsize(tfm)];
+       } sdesc;
+       char *kaddr;
+       int ret = 0;
+
+       sdesc.desc.tfm = tfm;
+       sdesc.desc.flags = 0;
+
+       kaddr = kmap_atomic(page);
+       ret = crypto_shash_digest(&sdesc.desc, kaddr, PAGE_SIZE,
+                                 out);
+       kunmap_atomic(kaddr);
+
+       return ret;
+}
+
+/*
+ * Return 1 when the extent already exists
+ * Return 0 when inserted the one into bytenr tree.
+ */
+static int insert_bytenr(struct rb_root *root, struct btrfs_dedup_hash *hash)
+{
+       struct rb_node **p = &root->rb_node;
+       struct rb_node *parent = NULL;
+       struct btrfs_dedup_hash *entry = NULL;
+
+       while (*p) {
+               parent = *p;
+               entry = rb_entry(parent, struct btrfs_dedup_hash,
+                                bytenr_node);
+               if (hash->bytenr + hash->offset < entry->bytenr + entry->offset)
+                       p = &(*p)->rb_left;
+               else if (hash->bytenr + hash->offset > entry->bytenr +
+                        entry->offset)
+                       p = &(*p)->rb_right;
+               else
+                       return 1;
+       }
+       rb_link_node(&hash->bytenr_node, parent, p);
+       rb_insert_color(&hash->bytenr_node, root);
+       return 0;
+}
+
+/*
+ * Must ensure insert_bytenr is called before, ensure this dedup_hash
+ * is not already in the tree
+ */
+static int insert_hash(struct rb_root *root, struct btrfs_dedup_hash *hash,
+                       int hash_len)
+{
+       struct rb_node **p = &root->rb_node;
+       struct rb_node *parent = NULL;
+       struct btrfs_dedup_hash *entry = NULL;
+
+       while (*p) {
+               parent = *p;
+               entry = rb_entry(parent, struct btrfs_dedup_hash,
+                                hash_node);
+               if (memcmp(hash->hash, entry->hash, hash_len) < 0)
+                       p = &(*p)->rb_left;
+               else if (memcmp(hash->hash, entry->hash, hash_len) > 0)
+                       p = &(*p)->rb_right;
+               /* Now hash matches, still allow it to be inserted */
+               else if (hash->bytenr + hash->offset < entry->bytenr +
+                        entry->offset)
+                       p = &(*p)->rb_left;
+               else if (hash->bytenr + hash->offset > entry->bytenr +
+                        entry->offset)
+                       p = &(*p)->rb_right;
+               else
+                       return 1;
+       }
+       rb_link_node(&hash->hash_node, parent, p);
+       rb_insert_color(&hash->hash_node, root);
+       return 0;
+}
+
+static int dedup_add_hash(struct btrfs_dedup_root *root,
+                         struct btrfs_dedup_hash *hash)
+{
+       int ret = 0;
+
+       WARN_ON(hash->bytenr == 0);
+
+       spin_lock(&root->lock);
+
+       ret = insert_bytenr(&root->bytenr_root, hash);
+       /* Already in tree */
+       if (ret > 0)
+               goto out;
+       insert_hash(&root->hash_root, hash,
+                   btrfs_dedup_sizes[root->dedup_type]);
+       list_add(&hash->lru_list, &root->lru_list);
+
+       root->current_nr++;
+
+       /* Remove the last dedup hash if we exceed limit */
+       while (root->limit_nr && root->current_nr > root->limit_nr) {
+               struct btrfs_dedup_hash *last;
+
+               last = list_entry(root->lru_list.prev, struct btrfs_dedup_hash,
+                                 lru_list);
+               __dedup_del_hash(root, last);
+               kmem_cache_free(btrfs_dedup_hash_cachep, last);
+       }
+out:
+       spin_unlock(&root->lock);
+       return ret;
+}
+
+/*
+ * Caller must hold lock on dedup_root->lock during the use of the hash.
+ * As the dedup_hash hash can be freed at anytime.
+ */
+static struct btrfs_dedup_hash *
+dedup_search_by_hash(struct btrfs_dedup_root *root, u8 *hash)
+{
+       struct rb_node **p = &root->hash_root.rb_node;
+       struct rb_node *parent = NULL;
+       struct btrfs_dedup_hash *entry = NULL;
+       int hash_len = btrfs_dedup_sizes[root->dedup_type];
+
+       while (*p) {
+               parent = *p;
+               entry = rb_entry(parent, struct btrfs_dedup_hash, hash_node);
+               if (memcmp(hash, entry->hash, hash_len) < 0)
+                       p = &(*p)->rb_left;
+               else if (memcmp(hash, entry->hash, hash_len) > 0)
+                       p = &(*p)->rb_right;
+               else {
+                       /* Found, need to re-add it to LRU list head */
+                       list_del(&entry->lru_list);
+                       list_add(&entry->lru_list, &root->lru_list);
+                       return entry;
+               }
+       }
+       return NULL;
+}
-- 
2.4.6

--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to