Hi guys, Have a nasty issue to report here. I have a RAID10 btrfs configuration where I was replacing one disk for another. Unfortunately, during the replace, it seems like one of my disks (not in the array) stopped getting powered (looks like the cable was loose) and caused the machine to lock up.
Since then, I have not been able to boot normally with the array in fstab. I’ve passed along “recovery” in my fstab line and have taken to editing in the fstab line to mount, then immediately commenting the line out (otherwise, the machine will not boot successfully; it’s a headless box, so this is painful). It seems to try to continue the replace, but dies shortly after mounting with kernel BUG at /home/kernel/COD/linux/fs/btrfs/volumes.c:5508 uname -a Linux ubuntu-server 4.4.0-040400rc6-generic #201512202030 SMP Mon Dec 21 01:32:09 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux btrfs --version btrfs-progs v4.0 btrfs fi show Label: none uuid: 48ed8a66-731d-499b-829e-dd07dd7260cc Total devices 9 FS bytes used 14.59TiB devid 0 size 4.55TiB used 3.16TiB path /dev/sde devid 1 size 4.55TiB used 3.16TiB path /dev/sdn devid 4 size 5.46TiB used 4.07TiB path /dev/sdh devid 5 size 5.46TiB used 4.07TiB path /dev/sdi devid 7 size 5.46TiB used 4.07TiB path /dev/sdm devid 8 size 5.46TiB used 4.07TiB path /dev/sdl devid 9 size 5.46TiB used 4.07TiB path /dev/sdg devid 10 size 5.46TiB used 4.07TiB path /dev/sdj devid 11 size 5.46TiB used 1.63TiB path /dev/sdk btrfs fi df /media/camino/ Data, RAID10: total=14.57TiB, used=14.57TiB System, RAID10: total=64.00MiB, used=1.28MiB Metadata, RAID10: total=25.47GiB, used=23.80GiB GlobalReserve, single: total=512.00MiB, used=0.00B dmesg output attached.
dmesg.log
Description: Binary data
Hope you guys have some ideas! Thanks, Asif