I have a home server using 2.6.15-vs2.1.0-gentoo-r1 kernel. It has several
raid arrays. The other day it froze, but I am away, so I had someone else
turn it off and on again. After that 2 of the arrays didn't come up. One
of them complained about invalid superblock on one partition but I added
that back and it synchronized and that work fine again.
However another array, md6, had 2 partitions complaining about invalid
superblock as shown below. It should use the partitions
/dev/hda8,/dev/hdc6,/dev/hdg7,/dev/hdh6,/dev/hde2
Two missing is one too many for raid 5 to just synchronize again. I have
looked around but cannot find any information about if or how this can be
fixed. Is it possible to fix the superblock or force it in some way. I
guess maybe they may be out of sync so the filesystem could be invalid if
forced to start anyway.
But maybe it would be possible to force it long enough to mount the
filesystem readonly for a while to retrieve those things not backed up
elsewhere and the recreate the array from scratch?
md: Autodetecting RAID arrays.
md: invalid raid superblock magic on hdg7
md: hdg7 has invalid sb, not importing!
md: invalid raid superblock magic on hdh6
md: hdh6 has invalid sb, not importing!
md: autorun ...
md: considering hdh7 ...
...
md: considering hde2 ...
md: adding hde2 ...
md: adding hdc6 ...
md: hdc2 has different UUID to hde2
md: hdc1 has different UUID to hde2
md: adding hda8 ...
md: hda2 has different UUID to hde2
md: hda1 has different UUID to hde2
md: created md6
md: bindhda8
md: bindhdc6
md: bindhde2
md: running: hde2hdc6hda8
md: md6: raid array is not clean -- starting background reconstruction
md: personality 4 is not loaded!
md: do_md_run() returned -22
md: md6 stopped.
md: unbindhde2
md: export_rdev(hde2)
md: unbindhdc6
md: export_rdev(hdc6)
md: unbindhda8
md: export_rdev(hda8)
...
--
Jesper
17:09:48 up 2 days, 59 min, 25 users, load average: 0.38, 0.88, 0.93
--
gentoo-user@gentoo.org mailing list