Re: MD or MDADM bug?
--Am Donnerstag, 1. September 2005 17:26 -0400 "David M. Strang" <[EMAIL PROTECTED]> schrieb: The problem is; my array is now 26 of 28 disks -- /dev/sdm *IS* bad; it [...] What can I do? I don't believe this is working as intended. I think the posts: 08.08.2005: How to recover a multiple raid5 disk failure with mdadm? 30.08.2005: 2 partition kicked from 6 raid5 describe the same problem. And it seems that no one was able to help. I hope you can rebuild your drive but I think you should use the backup if you need a quick solution. And indeed I think a "howto resolve a multiple raid5 disk failure" would be a good thing. Sometimes the problems are a faulty bus/cable and one knows that most (or even all) data is good. -- Claas Hilbrecht http://www.jucs-kramkiste.de - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: MD or MDADM bug?
--Am Donnerstag, 1. September 2005 17:26 -0400 "David M. Strang" <[EMAIL PROTECTED]> schrieb: The problem is; my array is now 26 of 28 disks -- /dev/sdm *IS* bad; it [...] What can I do? I don't believe this is working as intended. Maybe you didn't notice that but there are two recent threads that reports nearly the same problem. Look at the posts: 08.08.2005: How to recover a multiple raid5 disk failure with mdadm? 30.08.2005: 2 partition kicked from 6 raid5 And as far as I can see there is no solution yet. Maybe it's faster to restore the data from the backup instead of hoping someone can help you. But I really think a howto with "how to replace a multiple raid5 disk failure" is badly needed. Sometimes it happens that there is a bad cable or a problem on a resync. And the users knows that the data is good or at least most of the files are. -- Claas Hilbrecht http://www.jucs-kramkiste.de - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: [PATCH] proactive raid5 disk replacement for 2.6.11
--Am Sonntag, 14. August 2005 22:10 +0200 Pallai Roland <[EMAIL PROTECTED]> schrieb: this is a feature patch that implements 'proactive raid5 disk replacement' (http://www.arctic.org/~dean/raid-wishlist.html), After my experience with a broken raid5 (read the list) I think the "partially failed disks" feature you describe is really useful. I agree with you that this kind of error is rather common. -- Claas Hilbrecht http://www.jucs-kramkiste.de - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: How to recover a multiple raid5 disk failure with mdadm?
--Am Montag, 8. August 2005 09:18 +0200 Claas Hilbrecht <[EMAIL PROTECTED]> schrieb: The output below is from mdadm-2.0-devel-2 examine. What I don't understand is that there is difference in the "Spare Devices". Today I thought I could give the broken raid another try and use mdadm-2.0-devel-3. After looking again on the output of mdadm -E /dev/hde1 I think I had to repair this by hand. The Number Major Minor RaidDevice State this 3 3313 active sync /dev/hde1 0 0 00524288 spare 1 3670016 6553665536393216 spare 2 0 0131072589824 spare 3 2162688 65536196608393216 spare 4 3735552 655362621440 spare is totaly wrong. I think this is caused due to the bad checksum message in the mdadm -E output. Is there any way to fix this information with a disk editor or mdadm itself? -- Claas Hilbrecht http://www.jucs-kramkiste.de - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
How to recover a multiple raid5 disk failure with mdadm?
bd02:53de0ef9 Creation Time : Tue Dec 28 12:24:48 2004 Raid Level : raid5 Device Size : 80043136 (76.34 GiB 81.96 GB) Raid Devices : 4 Total Devices : 4 Preferred Minor : 0 Update Time : Sat Jul 23 20:23:19 2005 State : clean Active Devices : 2 Working Devices : 3 Failed Devices : 2 Spare Devices : 1 Checksum : c5646ffc - correct Events : 0.4340017 Layout : left-symmetric Chunk Size : 32K Number Major Minor RaidDevice State this 4 5714 spare /dev/hdk1 0 0 000 removed 1 1 5611 active sync /dev/hdi1 2 2 002 faulty removed 3 3 3313 active sync /dev/hde1 4 4 57 14 spare /dev/hdk1 ---***--- -- Claas Hilbrecht http://www.jucs-kramkiste.de - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html