New functions btrfs_qgroup_init/free_data_rsv_map() to init/free data reserve map.
Data reserve map is used to mark which range already holds reserved space, to avoid current reserved space leak. Signed-off-by: Qu Wenruo <quwen...@cn.fujitsu.com> --- fs/btrfs/btrfs_inode.h | 2 ++ fs/btrfs/inode.c | 10 +++++++ fs/btrfs/qgroup.c | 77 ++++++++++++++++++++++++++++++++++++++++++++++++++ fs/btrfs/qgroup.h | 3 ++ 4 files changed, 92 insertions(+) diff --git a/fs/btrfs/btrfs_inode.h b/fs/btrfs/btrfs_inode.h index e3ece65..27cc338 100644 --- a/fs/btrfs/btrfs_inode.h +++ b/fs/btrfs/btrfs_inode.h @@ -199,6 +199,8 @@ struct btrfs_inode { /* qgroup dirty map for data space reserve */ struct btrfs_qgroup_data_rsv_map *qgroup_rsv_map; + /* lock to ensure rsv_map will only be initialized once */ + spinlock_t qgroup_init_lock; }; extern unsigned char btrfs_filetype_table[]; diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c index 37dd8d0..61b2c17 100644 --- a/fs/btrfs/inode.c +++ b/fs/btrfs/inode.c @@ -8939,6 +8939,14 @@ struct inode *btrfs_alloc_inode(struct super_block *sb) INIT_LIST_HEAD(&ei->delalloc_inodes); RB_CLEAR_NODE(&ei->rb_node); + /* + * Init qgroup info to empty, as they will be initialized at write + * time. + * This behavior is needed for enable quota later case. + */ + spin_lock_init(&ei->qgroup_init_lock); + ei->qgroup_rsv_map = NULL; + return inode; } @@ -8996,6 +9004,8 @@ void btrfs_destroy_inode(struct inode *inode) btrfs_put_ordered_extent(ordered); } } + /* free and check data rsv map */ + btrfs_qgroup_free_data_rsv_map(inode); inode_tree_del(inode); btrfs_drop_extent_cache(inode, 0, (u64)-1, 0); free: diff --git a/fs/btrfs/qgroup.c b/fs/btrfs/qgroup.c index 561c36d..cf07c17 100644 --- a/fs/btrfs/qgroup.c +++ b/fs/btrfs/qgroup.c @@ -2539,3 +2539,80 @@ btrfs_qgroup_rescan_resume(struct btrfs_fs_info *fs_info) btrfs_queue_work(fs_info->qgroup_rescan_workers, &fs_info->qgroup_rescan_work); } + +/* + * Init data_rsv_map for a given inode. + * + * This is needed at write time as quota can be disabled and then enabled + */ +int btrfs_qgroup_init_data_rsv_map(struct inode *inode) +{ + struct btrfs_inode *binode = BTRFS_I(inode); + struct btrfs_root *root = binode->root; + struct btrfs_qgroup_data_rsv_map *dirty_map; + + if (!root->fs_info->quota_enabled || !is_fstree(root->objectid)) + return 0; + + spin_lock(&binode->qgroup_init_lock); + /* Quick route for init */ + if (likely(binode->qgroup_rsv_map)) + goto out; + spin_unlock(&binode->qgroup_init_lock); + + /* + * Slow allocation route + * + * TODO: Use kmem_cache to speedup allocation + */ + dirty_map = kmalloc(sizeof(*dirty_map), GFP_NOFS); + if (!dirty_map) + return -ENOMEM; + + dirty_map->reserved = 0; + dirty_map->root = RB_ROOT; + spin_lock_init(&dirty_map->lock); + + /* Lock again to ensure no one has already init it before */ + spin_lock(&binode->qgroup_init_lock); + if (binode->qgroup_rsv_map) { + spin_unlock(&binode->qgroup_init_lock); + kfree(dirty_map); + return 0; + } + binode->qgroup_rsv_map = dirty_map; +out: + spin_unlock(&binode->qgroup_init_lock); + return 0; +} + +void btrfs_qgroup_free_data_rsv_map(struct inode *inode) +{ + struct btrfs_inode *binode = BTRFS_I(inode); + struct btrfs_root *root = binode->root; + struct btrfs_qgroup_data_rsv_map *dirty_map = binode->qgroup_rsv_map; + struct rb_node *node; + + /* + * this function is called at inode destroy routine, so no concurrency + * will happen, no need to get the lock. + */ + if (!dirty_map) + return; + + /* insanity check */ + WARN_ON(!root->fs_info->quota_enabled || !is_fstree(root->objectid)); + + btrfs_qgroup_free(root, dirty_map->reserved); + spin_lock(&dirty_map->lock); + while ((node = rb_first(&dirty_map->root)) != NULL) { + struct data_rsv_range *range; + + range = rb_entry(node, struct data_rsv_range, node); + rb_erase(node, &dirty_map->root); + kfree(range); + } + spin_unlock(&dirty_map->lock); + kfree(dirty_map); + binode->qgroup_rsv_map = NULL; +} diff --git a/fs/btrfs/qgroup.h b/fs/btrfs/qgroup.h index 2f863a4..c87b7dc 100644 --- a/fs/btrfs/qgroup.h +++ b/fs/btrfs/qgroup.h @@ -84,4 +84,7 @@ int btrfs_verify_qgroup_counts(struct btrfs_fs_info *fs_info, u64 qgroupid, u64 rfer, u64 excl); #endif +/* for qgroup reserve */ +int btrfs_qgroup_init_data_rsv_map(struct inode *inode); +void btrfs_qgroup_free_data_rsv_map(struct inode *inode); #endif /* __BTRFS_QGROUP__ */ -- 2.5.1 -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html