Re: MD or MDADM bug?

2005-09-01 Thread Claas Hilbrecht
--Am Donnerstag, 1. September 2005 17:26 -0400 "David M. Strang" 
<[EMAIL PROTECTED]> schrieb:



The problem is; my array is now 26 of 28 disks -- /dev/sdm *IS* bad; it

[...]

What can I do? I don't believe this is working as intended.


I think the posts:

08.08.2005: How to recover a multiple raid5 disk failure with mdadm?
30.08.2005: 2 partition kicked from 6 raid5

describe the same problem. And it seems that no one was able to help. I 
hope you can rebuild your drive but I think you should use the backup if 
you need a quick solution. And indeed I think a "howto resolve a multiple 
raid5 disk failure" would be a good thing. Sometimes the problems are a 
faulty bus/cable and one knows that most (or even all) data is good.


--
Claas Hilbrecht
http://www.jucs-kramkiste.de


-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: MD or MDADM bug?

2005-09-01 Thread Claas Hilbrecht
--Am Donnerstag, 1. September 2005 17:26 -0400 "David M. Strang" 
<[EMAIL PROTECTED]> schrieb:



The problem is; my array is now 26 of 28 disks -- /dev/sdm *IS* bad; it

[...]

What can I do? I don't believe this is working as intended.


Maybe you didn't notice that but there are two recent threads that reports 
nearly the same problem. Look at the posts:


08.08.2005: How to recover a multiple raid5 disk failure with mdadm?
30.08.2005: 2 partition kicked from 6 raid5

And as far as I can see there is no solution yet. Maybe it's faster to 
restore the data from the backup instead of hoping someone can help you. 
But I really think a howto with "how to replace a multiple raid5 disk 
failure" is badly needed. Sometimes it happens that there is a bad cable or 
a problem on a resync. And the users knows that the data is good or at 
least most of the files are.


--
Claas Hilbrecht
http://www.jucs-kramkiste.de


-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: [PATCH] proactive raid5 disk replacement for 2.6.11

2005-08-14 Thread Claas Hilbrecht
--Am Sonntag, 14. August 2005 22:10 +0200 Pallai Roland <[EMAIL PROTECTED]> 
schrieb:



 this is a feature patch that implements 'proactive raid5 disk
replacement' (http://www.arctic.org/~dean/raid-wishlist.html),


After my experience with a broken raid5 (read the list) I think the 
"partially failed disks" feature you describe is really useful. I agree 
with you that this kind of error is rather common.


--
Claas Hilbrecht
http://www.jucs-kramkiste.de

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: How to recover a multiple raid5 disk failure with mdadm?

2005-08-11 Thread Claas Hilbrecht
--Am Montag, 8. August 2005 09:18 +0200 Claas Hilbrecht 
<[EMAIL PROTECTED]> schrieb:



The output below is from mdadm-2.0-devel-2 examine. What I don't
understand is that there is difference in the "Spare Devices".


Today I thought I could give the broken raid another try and use 
mdadm-2.0-devel-3. After looking again on the output of mdadm -E /dev/hde1 
I think I had to repair this by hand. The


 Number   Major   Minor   RaidDevice State
this 3  3313  active sync   /dev/hde1

  0 0   00524288  spare
  1 3670016   6553665536393216  spare
  2 0   0131072589824  spare
  3 2162688   65536196608393216  spare
  4 3735552   655362621440  spare

is totaly wrong. I think this is caused due to the bad checksum message in 
the mdadm -E output. Is there any way to fix this information with a disk 
editor or mdadm itself?


--
Claas Hilbrecht
http://www.jucs-kramkiste.de

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


How to recover a multiple raid5 disk failure with mdadm?

2005-08-08 Thread Claas Hilbrecht
bd02:53de0ef9
 Creation Time : Tue Dec 28 12:24:48 2004
Raid Level : raid5
   Device Size : 80043136 (76.34 GiB 81.96 GB)
  Raid Devices : 4
 Total Devices : 4
Preferred Minor : 0

   Update Time : Sat Jul 23 20:23:19 2005
 State : clean
Active Devices : 2
Working Devices : 3
Failed Devices : 2
 Spare Devices : 1
  Checksum : c5646ffc - correct
Events : 0.4340017

Layout : left-symmetric
Chunk Size : 32K

 Number   Major   Minor   RaidDevice State
this 4  5714  spare   /dev/hdk1

  0 0   000  removed
  1 1  5611  active sync   /dev/hdi1
  2 2   002  faulty removed
  3 3  3313  active sync   /dev/hde1
  4 4  57    14  spare   /dev/hdk1
---***---

--
Claas Hilbrecht
http://www.jucs-kramkiste.de

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html