On 2019/2/27 下午8:22, David Sterba wrote: > On Sat, Feb 23, 2019 at 08:47:09AM +0800, Qu Wenruo wrote: >> >> >> On 2019/2/22 下午11:18, David Sterba wrote: >>> On Mon, Feb 18, 2019 at 01:27:41PM +0800, Qu Wenruo wrote: >>>> v5.1: >>>> - Add "block=%llu " output for write/read time error line. >>>> - Also output read time error message for fsid/start/level check. >>> >>> I see reports from tests btrfs/124 and btrfs/125, >> >> That two tests are RAID1 reconstruction. >> It's known to trigger read error on missing devices. >> >> Especially for bytenr mismatch, as all read from missing device will >> result 0 filled page. >> >> I could change the read time error message to skip bytenr/csum mismatch >> to avoid such report. > > Yeah the output needs some tuning. See the example below. >> [snip] >>> [ 2860.244462] BTRFS error (device vdb): bad tree block start, want >>> 31703040 have 0 >>> [ 2860.248013] BTRFS error (device vdb): block=31703040 read time tree >>> block corruption detected >>> [ 2860.284017] BTRFS error (device vdb): block=31719424 read time tree >>> block corruption detected >>> [ 2860.315705] BTRFS error (device vdb): block=31735808 read time tree >>> block corruption detected >>> [ 2860.336001] BTRFS error (device vdb): block=31752192 read time tree >>> block corruption detected > > Block 31752192 does not have any previous message about a specific error > (unlike the 'bad tree block' messages of other blocks). I don't think > it's good to print just the 'read time' message.
The first is caused by the bytenr mismatch. But for the remaining 3, they can be caused by read error from missing device. EXTENT_BUFFER_READ_ERR would cause no detailed error report here. > >>> [ 2860.336054] BTRFS error (device vdb): block=31768576 read time tree >>> block corruption detected >>> [ 2860.339786] BTRFS error (device vdb): block=31784960 read time tree >>> block corruption detected >>> [ 2860.366693] BTRFS error (device vdb): block=31768576 read time tree >>> block corruption detected >>> [ 2860.389495] BTRFS error (device vdb): block=31784960 read time tree >>> block corruption detected >>> [ 2860.418551] BTRFS error (device vdb): block=31866880 read time tree >>> block corruption detected >>> [ 2860.418713] BTRFS error (device vdb): block=31883264 read time tree >>> block corruption detected >>> [ 2860.418721] BTRFS error (device vdb): block=31916032 read time tree >>> block corruption detected >>> [ 2860.418736] BTRFS error (device vdb): block=31932416 read time tree >>> block corruption detected >>> [ 2860.418744] BTRFS error (device vdb): block=31948800 read time tree >>> block corruption detected >>> [ 2860.419721] BTRFS error (device vdb): block=31899648 read time tree >>> block corruption detected >>> [ 2860.462401] BTRFS error (device vdb): block=31866880 read time tree >>> block corruption detected >>> [ 2860.490821] BTRFS error (device vdb): block=31932416 read time tree >>> block corruption detected >>> [ 2860.506718] BTRFS error (device vdb): block=31899648 read time tree >>> block corruption detected >>> [ 2860.535013] BTRFS error (device vdb): block=31916032 read time tree >>> block corruption detected >>> [ 2860.568602] BTRFS error (device vdb): block=31948800 read time tree >>> block corruption detected >>> [ 2860.590434] BTRFS error (device vdb): block=31965184 read time tree >>> block corruption detected >>> [ 2860.590532] BTRFS error (device vdb): block=31981568 read time tree >>> block corruption detected >>> [ 2860.590677] BTRFS error (device vdb): block=31997952 read time tree >>> block corruption detected >>> [ 2860.590688] BTRFS error (device vdb): block=32014336 read time tree >>> block corruption detected >>> >>> [ 2860.590697] BTRFS error (device vdb): block=32030720 read time tree >>> block corruption detected >>> [ 2860.590705] BTRFS error (device vdb): block=32047104 read time tree >>> block corruption detected >>> [ 2860.625834] BTRFS error (device vdb): block=31965184 read time tree >>> block corruption detected >>> [ 2860.658723] BTRFS error (device vdb): block=31981568 read time tree >>> block corruption detected >>> [ 2860.696644] BTRFS error (device vdb): block=32014336 read time tree >>> block corruption detected >>> [ 2860.735014] BTRFS error (device vdb): block=32030720 read time tree >>> block corruption detected >>> [ 2860.743240] BTRFS error (device vdb): block=32096256 read time tree >>> block corruption detected >>> [ 2860.780296] BTRFS error (device vdb): block=32047104 read time tree >>> block corruption detected >>> [ 2860.807073] BTRFS error (device vdb): block=32063488 read time tree >>> block corruption detected >>> [ 2860.807173] BTRFS error (device vdb): block=32079872 read time tree >>> block corruption detected >>> [ 2860.841582] BTRFS error (device vdb): block=32079872 read time tree >>> block corruption detected > > This long sequence of messages lacks details about what happen, and > there are duplicates (block 32079872). If this is caused by reads from > missing device, this can potentially flood the logs. Not only from missing device, but also from stale old device. Anyway, since I have reverted the behavior to the original one, which only output the "read time" error message for tree checker error, it should no longer flood the kernel message. Thanks, Qu >
signature.asc
Description: OpenPGP digital signature