Hi again, the check went through, and no errors were reported...
I tried mounting the fs again, and suddenly all read errors disappeared! A full copy is already running, and after that i will nuke it from orbit. Thank you all for the quick answers, I do not completely understand what exactly was going wrong, but I am confident that the files are not corrupted anymore (I checked a few random ones). -- Daniel (output of the check) # btrfs check --readonly /dev/mapper/sde-open Opening filesystem to check... Checking filesystem on /dev/mapper/sde-open UUID: 26a3ef79-15ba-4041-951b-284fbbe08074 [1/7] checking root items [2/7] checking extents [3/7] checking free space cache [4/7] checking fs roots [5/7] checking only csums items (without verifying data) [6/7] checking root refs [7/7] checking quota groups skipped (not enabled on this FS) found 50094867738635 bytes used, no error found total csum bytes: 48133927932 total tree bytes: 60644540416 total fs tree bytes: 4307402752 total extent tree bytes: 1874182144 btree space waste bytes: 6642842911 file data blocks allocated: 50464303362048 referenced 49273323020288 Am Di., 16. Apr. 2019 um 10:00 Uhr schrieb Roman Mamedov <r...@romanrm.net>: > > On Tue, 16 Apr 2019 09:46:39 +0200 > Daniel Brunner <daniel@brunner.ninja> wrote: > > > Hi, > > > > thanks for the quick response. > > > > The filesystem went read-only on its own right at the first read error. > > I unmounted all mounts and rebooted (just to be sure). > > > > I ran the command you suggested with --progress > > All output is flushed away with thousands of lines like those at the > > end of the log paste. > > > > Does it make sense to let it run until the end or can I assume that 2 > > drives are bad? > > Also I checked SMART values but they seem to be ok. > > > > https://0x0.st/zNg6.log > > ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED > WHEN_FAILED RAW_VALUE > 190 Airflow_Temperature_Cel 0x0022 047 031 040 Old_age Always > In_the_past 53 (Min/Max 47/53 #5115) > > This seems to be normalized as VALUE=100-RAW_VALUE by the SMART firmware, and > looking at the reading in WORST, indicates that some of your drives earlier > have seen temperatures of as high as 69 C. > > This is insanely hot to run your drives at, I'd say to the point of "shut off > everything ASAP via the mains breaker to avoid immediate permanent damage"; > > Not sure if it's related to the csum errors at hand, but it very well might > be. > > Even the current temps of 55-60 are about 15-20 degrees higher than ideal. > > -- > With respect, > Roman