New functions btrfs_qgroup_init/free_data_rsv_map() to init/free data
reserve map.

Data reserve map is used to mark which range already holds reserved
space, to avoid current reserved space leak.

Signed-off-by: Qu Wenruo <quwen...@cn.fujitsu.com>
---
 fs/btrfs/btrfs_inode.h |  2 ++
 fs/btrfs/inode.c       | 10 +++++++
 fs/btrfs/qgroup.c      | 77 ++++++++++++++++++++++++++++++++++++++++++++++++++
 fs/btrfs/qgroup.h      |  3 ++
 4 files changed, 92 insertions(+)

diff --git a/fs/btrfs/btrfs_inode.h b/fs/btrfs/btrfs_inode.h
index e3ece65..27cc338 100644
--- a/fs/btrfs/btrfs_inode.h
+++ b/fs/btrfs/btrfs_inode.h
@@ -199,6 +199,8 @@ struct btrfs_inode {
 
        /* qgroup dirty map for data space reserve */
        struct btrfs_qgroup_data_rsv_map *qgroup_rsv_map;
+       /* lock to ensure rsv_map will only be initialized once */
+       spinlock_t qgroup_init_lock;
 };
 
 extern unsigned char btrfs_filetype_table[];
diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c
index 37dd8d0..61b2c17 100644
--- a/fs/btrfs/inode.c
+++ b/fs/btrfs/inode.c
@@ -8939,6 +8939,14 @@ struct inode *btrfs_alloc_inode(struct super_block *sb)
        INIT_LIST_HEAD(&ei->delalloc_inodes);
        RB_CLEAR_NODE(&ei->rb_node);
 
+       /*
+        * Init qgroup info to empty, as they will be initialized at write
+        * time.
+        * This behavior is needed for enable quota later case.
+        */
+       spin_lock_init(&ei->qgroup_init_lock);
+       ei->qgroup_rsv_map = NULL;
+
        return inode;
 }
 
@@ -8996,6 +9004,8 @@ void btrfs_destroy_inode(struct inode *inode)
                        btrfs_put_ordered_extent(ordered);
                }
        }
+       /* free and check data rsv map */
+       btrfs_qgroup_free_data_rsv_map(inode);
        inode_tree_del(inode);
        btrfs_drop_extent_cache(inode, 0, (u64)-1, 0);
 free:
diff --git a/fs/btrfs/qgroup.c b/fs/btrfs/qgroup.c
index 561c36d..cf07c17 100644
--- a/fs/btrfs/qgroup.c
+++ b/fs/btrfs/qgroup.c
@@ -2539,3 +2539,80 @@ btrfs_qgroup_rescan_resume(struct btrfs_fs_info *fs_info)
                btrfs_queue_work(fs_info->qgroup_rescan_workers,
                                 &fs_info->qgroup_rescan_work);
 }
+
+/*
+ * Init data_rsv_map for a given inode.
+ *
+ * This is needed at write time as quota can be disabled and then enabled
+ */
+int btrfs_qgroup_init_data_rsv_map(struct inode *inode)
+{
+       struct btrfs_inode *binode = BTRFS_I(inode);
+       struct btrfs_root *root = binode->root;
+       struct btrfs_qgroup_data_rsv_map *dirty_map;
+
+       if (!root->fs_info->quota_enabled || !is_fstree(root->objectid))
+               return 0;
+
+       spin_lock(&binode->qgroup_init_lock);
+       /* Quick route for init */
+       if (likely(binode->qgroup_rsv_map))
+               goto out;
+       spin_unlock(&binode->qgroup_init_lock);
+
+       /*
+        * Slow allocation route
+        *
+        * TODO: Use kmem_cache to speedup allocation
+        */
+       dirty_map = kmalloc(sizeof(*dirty_map), GFP_NOFS);
+       if (!dirty_map)
+               return -ENOMEM;
+
+       dirty_map->reserved = 0;
+       dirty_map->root = RB_ROOT;
+       spin_lock_init(&dirty_map->lock);
+
+       /* Lock again to ensure no one has already init it before */
+       spin_lock(&binode->qgroup_init_lock);
+       if (binode->qgroup_rsv_map) {
+               spin_unlock(&binode->qgroup_init_lock);
+               kfree(dirty_map);
+               return 0;
+       }
+       binode->qgroup_rsv_map = dirty_map;
+out:
+       spin_unlock(&binode->qgroup_init_lock);
+       return 0;
+}
+
+void btrfs_qgroup_free_data_rsv_map(struct inode *inode)
+{
+       struct btrfs_inode *binode = BTRFS_I(inode);
+       struct btrfs_root *root = binode->root;
+       struct btrfs_qgroup_data_rsv_map *dirty_map = binode->qgroup_rsv_map;
+       struct rb_node *node;
+
+       /*
+        * this function is called at inode destroy routine, so no concurrency
+        * will happen, no need to get the lock.
+        */
+       if (!dirty_map)
+               return;
+
+       /* insanity check */
+       WARN_ON(!root->fs_info->quota_enabled || !is_fstree(root->objectid));
+
+       btrfs_qgroup_free(root, dirty_map->reserved);
+       spin_lock(&dirty_map->lock);
+       while ((node = rb_first(&dirty_map->root)) != NULL) {
+               struct data_rsv_range *range;
+
+               range = rb_entry(node, struct data_rsv_range, node);
+               rb_erase(node, &dirty_map->root);
+               kfree(range);
+       }
+       spin_unlock(&dirty_map->lock);
+       kfree(dirty_map);
+       binode->qgroup_rsv_map = NULL;
+}
diff --git a/fs/btrfs/qgroup.h b/fs/btrfs/qgroup.h
index 2f863a4..c87b7dc 100644
--- a/fs/btrfs/qgroup.h
+++ b/fs/btrfs/qgroup.h
@@ -84,4 +84,7 @@ int btrfs_verify_qgroup_counts(struct btrfs_fs_info *fs_info, 
u64 qgroupid,
                               u64 rfer, u64 excl);
 #endif
 
+/* for qgroup reserve */
+int btrfs_qgroup_init_data_rsv_map(struct inode *inode);
+void btrfs_qgroup_free_data_rsv_map(struct inode *inode);
 #endif /* __BTRFS_QGROUP__ */
-- 
2.5.1

--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to