At 03/30/2017 08:51 AM, Liu Bo wrote:
On Wed, Mar 29, 2017 at 09:33:19AM +0800, Qu Wenruo wrote:
[...]

Reported-by: Goffredo Baroncelli <[email protected]>
Signed-off-by: Qu Wenruo <[email protected]>
---
 fs/btrfs/scrub.c | 14 ++++++++++++++
 1 file changed, 14 insertions(+)

diff --git a/fs/btrfs/scrub.c b/fs/btrfs/scrub.c
index ab33b9a8aac2..f92d2512f4f3 100644
--- a/fs/btrfs/scrub.c
+++ b/fs/btrfs/scrub.c
@@ -1129,6 +1129,17 @@ static int scrub_handle_errored_block(struct scrub_block 
*sblock_to_check)
        have_csum = sblock_to_check->pagev[0]->have_csum;
        dev = sblock_to_check->pagev[0]->dev;

+       /*
+        * For RAID5/6 race can happen for different dev scrub thread.
+        * For data corruption, Parity and Data thread will both try
+        * to recovery the data.
+        * Race can lead to double added csum error, or even unrecoverable
+        * error.
+        */
+       ret = lock_full_stripe(fs_info, logical);
+       if (ret < 0)
+               return ret;
+

Firstly, sctx->stat needs to be set with errors before returning errors.

Right.

I'll update malloc_errors for ret == -ENOMEM case and uncorrectable_errors for ret == -ENOENT case.


Secondly, I think the critical section starts right before re-checking the
failed mirror, doesn't it?  If so, we could get the benefit from
sblock_bad->page since the page->recover can tell if this is a raid56 profile so
that we may save a searching for block group.

It's true we can save a block group search, but that's relying on bbio->flags.

I prefer to make lock/unlock_full_stripe() to be as independent as possible, so such modification is not favoured.


Although I'm looking forward better scrub structure which allow us to get block group cache easier. If we can get bg_cache easier in such context, then lock/unlock_full_stripe() have no need to search bg_cache by themselves.

But for now, I prefer to trade a little performance for more independent code.


        if (sctx->is_dev_replace && !is_metadata && !have_csum) {
                sblocks_for_recheck = NULL;
                goto nodatasum_case;
@@ -1463,6 +1474,9 @@ static int scrub_handle_errored_block(struct scrub_block 
*sblock_to_check)
                kfree(sblocks_for_recheck);
        }

+       ret = unlock_full_stripe(fs_info, logical);
+       if (ret < 0)
+               return ret;

Seems that the callers of scrub_handle_errored_block doesn't care about @ret.

And could you please put a 'locked' flag after taking the lock succesfully?
Otherwise every raid profile has to check block group for raid flag.

I'm OK to introduce a bool as new paramter for lock/unlock_full_stripe() as a shortcut to avoid bg search at unlock.

But the true fix would be scrub structure cleanup to make us accessing bg_cache without hassle.

Thanks for the review,
Qu


Thanks,

-liubo
        return 0;
 }

--
2.12.1



--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html




--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to