-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA1 Alex Samad wrote: > On Sun, Jun 22, 2008 at 12:28:03PM -0400, Matt Gracie wrote: > >> [snip] > > The problem is that when I tried, using "mdadm /dev/md0 --add > /dev/sdd1", the rebuild would kick off and then fail after a short time, > marking all four drives as faulty. > > So I rebooted, running the RAID in degraded mode, fdisked and mkfsed > /dev/sdd1 again, and tried again. Same result. > > I replaced the firewire PCI card. Same result. > > >> [snip] > > Does anyone have any suggestions for cleaning up my degraded RAID set so > that I can put another drive in the pool? All of my data seems to be > okay (although you'd better believe I've backed the irreplaceables up > again to be sure), but I really don't want to deal with another disk > failure. > >> sounds like you have done all the things I would have done. > >> can you post a cat /proc/mdstat
mogwai:~# cat /proc/mdstat Personalities : [raid6] [raid5] [raid4] md0 : active raid5 sdb1[0] sdc1[2] sda1[1] 732587712 blocks level 5, 64k chunk, algorithm 2 [4/3] [UUU_] unused devices: <none> > >> and a > >> mdadm --query --detail /dev/md0 mogwai:~# mdadm --query --detail /dev/md0 /dev/md0: Version : 00.90.03 Creation Time : Sun Mar 30 14:12:22 2008 Raid Level : raid5 Array Size : 732587712 (698.65 GiB 750.17 GB) Used Dev Size : 244195904 (232.88 GiB 250.06 GB) Raid Devices : 4 Total Devices : 3 Preferred Minor : 0 Persistence : Superblock is persistent Update Time : Mon Jun 23 18:10:12 2008 State : clean, degraded Active Devices : 3 Working Devices : 3 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 64K UUID : 36c5a427:cb5a2218:4a8af7ff:69db778e Events : 0.13586 Number Major Minor RaidDevice State 0 8 17 0 active sync /dev/sdb1 1 8 1 1 active sync /dev/sda1 2 8 33 2 active sync /dev/sdc1 3 0 0 3 removed The problem (where all the drives are marked faulty) seems to be a result of large amounts of activity on the FW bus; when I attempted to back up some stuff from the degraded RAID set onto another firewire drive, the RAID failed and everything was marked faulty. When I backed up the same directory onto an internal IDE drive, however, there was no problem. So, I think this might be somehow specific to the interaction between software RAID and the Firewire subsystem. For the record: mogwai:~# uname -a Linux mogwai 2.6.25-2-686 #1 SMP Tue May 27 15:38:35 UTC 2008 i686 GNU/Linux - --Matt -----BEGIN PGP SIGNATURE----- Version: GnuPG v1.4.9 (GNU/Linux) Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org iEYEARECAAYFAkhgalQACgkQ6wBqCH7pnjIhUACfeUOgAKKG/J7ugwHWfmf4vqaU HfQAoN4apPeQhDx9Uj7Dm7MizALFlGc8 =GfBQ -----END PGP SIGNATURE----- -- To UNSUBSCRIBE, email to [EMAIL PROTECTED] with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]