On Thu, Feb 2, 2017 at 6:53 PM, Liu Bo <bo.li....@oracle.com> wrote: > On Thu, Feb 02, 2017 at 06:32:01PM +0000, Filipe Manana wrote: >> On Thu, Feb 2, 2017 at 6:19 PM, Liu Bo <bo.li....@oracle.com> wrote: >> > On Wed, Feb 01, 2017 at 11:01:28PM +0000, fdman...@kernel.org wrote: >> >> From: Filipe Manana <fdman...@suse.com> >> >> >> >> At close_ctree() we free the block groups and then only after we wait for >> >> any running worker kthreads to finish and shutdown the workqueues. This >> >> behaviour is racy and it triggers an assertion failure when freeing block >> >> groups because while we are doing it we can have for example a block group >> >> caching kthread running, and in that case the block group's reference >> >> count is greater than 1, leading to an assertion failure: >> >> >> >> [19041.198004] assertion failed: atomic_read(&block_group->count) == 1, >> >> file: fs/btrfs/extent-tree.c, line: 9799 >> >> [19041.200584] ------------[ cut here ]------------ >> >> [19041.201692] kernel BUG at fs/btrfs/ctree.h:3418! >> >> [19041.202830] invalid opcode: 0000 [#1] PREEMPT SMP >> >> [19041.203929] Modules linked in: btrfs xor raid6_pq dm_flakey dm_mod >> >> crc32c_generic ppdev sg psmouse acpi_cpufreq pcspkr parport_pc evdev >> >> tpm_tis parport tpm_tis_core i2c_piix4 i2c_core tpm serio_raw processor >> >> button loop autofs4 ext4 crc16 jbd2 mbcache sr_mod cdrom sd_mod >> >> ata_generic virtio_scsi ata_piix virtio_pci libata virtio_ring virtio >> >> e1000 scsi_mod floppy [last unloaded: btrfs] >> >> [19041.208082] CPU: 6 PID: 29051 Comm: umount Not tainted >> >> 4.9.0-rc7-btrfs-next-36+ #1 >> >> [19041.208082] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), >> >> BIOS rel-1.9.1-0-gb3ef39f-prebuilt.qemu-project.org 04/01/2014 >> >> [19041.208082] task: ffff88015f028980 task.stack: ffffc9000ad34000 >> >> [19041.208082] RIP: 0010:[<ffffffffa03e319e>] [<ffffffffa03e319e>] >> >> assfail.constprop.41+0x1c/0x1e [btrfs] >> >> [19041.208082] RSP: 0018:ffffc9000ad37d60 EFLAGS: 00010286 >> >> [19041.208082] RAX: 0000000000000061 RBX: ffff88015ecb4000 RCX: >> >> 0000000000000001 >> >> [19041.208082] RDX: ffff88023f392fb8 RSI: ffffffff817ef7ba RDI: >> >> 00000000ffffffff >> >> [19041.208082] RBP: ffffc9000ad37d60 R08: 0000000000000001 R09: >> >> 0000000000000000 >> >> [19041.208082] R10: ffffc9000ad37cb0 R11: ffffffff82f2b66d R12: >> >> ffff88023431d170 >> >> [19041.208082] R13: ffff88015ecb40c0 R14: ffff88023431d000 R15: >> >> ffff88015ecb4100 >> >> [19041.208082] FS: 00007f44f3d42840(0000) GS:ffff88023f380000(0000) >> >> knlGS:0000000000000000 >> >> [19041.208082] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 >> >> [19041.208082] CR2: 00007f65d623b000 CR3: 00000002166f2000 CR4: >> >> 00000000000006e0 >> >> [19041.208082] Stack: >> >> [19041.208082] ffffc9000ad37d98 ffffffffa035989f ffff88015ecb4000 >> >> ffff88015ecb5630 >> >> [19041.208082] ffff88014f6be000 0000000000000000 00007ffcf0ba6a10 >> >> ffffc9000ad37df8 >> >> [19041.208082] ffffffffa0368cd4 ffff88014e9658e0 ffffc9000ad37e08 >> >> ffffffff811a634d >> >> [19041.208082] Call Trace: >> >> [19041.208082] [<ffffffffa035989f>] btrfs_free_block_groups+0x17f/0x392 >> >> [btrfs] >> >> [19041.208082] [<ffffffffa0368cd4>] close_ctree+0x1c5/0x2e1 [btrfs] >> >> [19041.208082] [<ffffffff811a634d>] ? evict_inodes+0x132/0x141 >> >> [19041.208082] [<ffffffffa034356d>] btrfs_put_super+0x15/0x17 [btrfs] >> >> [19041.208082] [<ffffffff8118fc32>] generic_shutdown_super+0x6a/0xeb >> >> [19041.208082] [<ffffffff8119004f>] kill_anon_super+0x12/0x1c >> >> [19041.208082] [<ffffffffa0343370>] btrfs_kill_super+0x16/0x21 [btrfs] >> >> [19041.208082] [<ffffffff8118fad1>] deactivate_locked_super+0x3b/0x68 >> >> [19041.208082] [<ffffffff8118fb34>] deactivate_super+0x36/0x39 >> >> [19041.208082] [<ffffffff811a9946>] cleanup_mnt+0x58/0x76 >> >> [19041.208082] [<ffffffff811a99a2>] __cleanup_mnt+0x12/0x14 >> >> [19041.208082] [<ffffffff81071573>] task_work_run+0x6f/0x95 >> >> [19041.208082] [<ffffffff81001897>] prepare_exit_to_usermode+0xa3/0xc1 >> >> [19041.208082] [<ffffffff81001a23>] syscall_return_slowpath+0x16e/0x1d2 >> >> [19041.208082] [<ffffffff814c607d>] entry_SYSCALL_64_fastpath+0xab/0xad >> >> [19041.208082] Code: c7 ae a0 3e a0 48 89 e5 e8 4e 74 d4 e0 0f 0b 55 89 >> >> f1 48 c7 c2 0b a4 3e a0 48 89 fe 48 c7 c7 a4 a6 3e a0 48 89 e5 e8 30 74 >> >> d4 e0 <0f> 0b 55 31 d2 48 89 e5 e8 d5 b9 f7 ff 5d c3 48 63 f6 55 31 c9 >> >> [19041.208082] RIP [<ffffffffa03e319e>] assfail.constprop.41+0x1c/0x1e >> >> [btrfs] >> >> [19041.208082] RSP <ffffc9000ad37d60> >> >> [19041.279264] ---[ end trace 23330586f16f064d ]--- >> >> >> >> This started happening as of kernel 4.8, since commit f3bca8028bd9 >> >> ("Btrfs: add ASSERT for block group's memory leak") introduced these >> >> assertions. >> >> >> >> So fix this by freeing the block groups only after waiting for all >> >> worker kthreads to complete and shutdown the workqueues. >> > >> > This looks good to me, but I don't understand how that could happen, if >> > a block group is being cached by the caching worker thread, the block >> > group cache has been marked as BTRFS_CACHE_STARTED so we should wait on >> > wait_block_group_cache_done() in btrfs_free_block_groups() before >> > getting to the ASSERT. Maybe something else broke? >> >> Simple. Look at extent-tree.c:caching_kthread() - the the caching > > caching_thread() vs caching_kthread(), free space cache vs inode cache, > confusing helper names... > >> state is updated (to error or finished), but only much later (at the >> very end) the kthread drops its ref count on the block group. So the >> assertion you added in commit f3bca8028bd9 fails because the block >> group's ref count is 2 and not 1. > > I see. > > It doens't make sense to load cache when we're closing the FS, and > looks like it's not necessary to put btrfs_free_block_group at the very > end of caching_thread(), we could free it before waking up waiters.
It would still be racy. The task calling free_block_groups() could just have seen the caching state set to finished/error just after the caching kthread set it and before it unlocked the block group's spinlock or before/while it calls free_excluded_extents() for example. Iow, make the wakeup after the call to put would still not make it safe. > >> So nothing broke, just the assertion made an incorrect assumption. But >> I think it's good having that and the other assertions in place to >> detect issues, that's why the patch doesn't remove it/them and makes >> them safe instead. >> > > +1. > > Thanks, > > -liubo > >> > >> > Thanks, >> > >> > -liubo >> >> >> >> Signed-off-by: Filipe Manana <fdman...@suse.com> >> >> --- >> >> fs/btrfs/disk-io.c | 9 +++++++-- >> >> 1 file changed, 7 insertions(+), 2 deletions(-) >> >> >> >> diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c >> >> index 066d9b9..a90e40e 100644 >> >> --- a/fs/btrfs/disk-io.c >> >> +++ b/fs/btrfs/disk-io.c >> >> @@ -3985,8 +3985,6 @@ void close_ctree(struct btrfs_fs_info *fs_info) >> >> >> >> btrfs_put_block_group_cache(fs_info); >> >> >> >> - btrfs_free_block_groups(fs_info); >> >> - >> >> /* >> >> * we must make sure there is not any read request to >> >> * submit after we stopping all workers. >> >> @@ -3994,6 +3992,13 @@ void close_ctree(struct btrfs_fs_info *fs_info) >> >> invalidate_inode_pages2(fs_info->btree_inode->i_mapping); >> >> btrfs_stop_all_workers(fs_info); >> >> >> >> + /* >> >> + * Free block groups only after stopping all workers, since we could >> >> + * have block group caching kthreads running, and therefore they >> >> could >> >> + * race with us if we freed the block groups before stopping them. >> >> + */ >> >> + btrfs_free_block_groups(fs_info); >> >> + >> >> clear_bit(BTRFS_FS_OPEN, &fs_info->flags); >> >> free_root_pointers(fs_info, 1); >> >> >> >> -- >> >> 2.7.0.rc3 >> >> >> >> -- >> >> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in >> >> the body of a message to majord...@vger.kernel.org >> >> More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html