On 06/19/2012 09:13 PM, Michael Butash wrote:
So yeah, no raid is perfect...

I've been using software raid1 (md) for a while now for my desktops and
laptops work and home, and since my adventures in ati gpu land, I've
twice now had video software/hardware cause my software raid to fail
ugly, but both times survivable while I rebuilt the array manually. This
was just a few days ago the last...

Both times were using GL functions (this time toggling compositing
on/off, last time i think minecraft) that caused the ati fglrx drivers
to spew hardware errors seeming to glitch the card itself.  Two separate
cards as well now.  Getting back into desktop went into visa with gpu
unavailable.  Then I saw my raid was degraded, again, same timestamp as
the gpu glitch.

First time prior one of the two disks in the md for boot went offline,
simply added sdb1/2 back.  This time one partition on each disk to the
two md's (boot/else) to go offline alternatively (sda1/sdb2) - very odd.
  The second disk wouldn't respond to hdparm/fdisk query until a reboot
that was done very hesitantly and not before I backed up anything I
cared about to an nfs share.  Data on both remained available which was
really the odd part.

To its testament, it rebooted, both disks reported healthy (hdparm,
ubuntu disk utility), I re-added each partition, let it rebuild, and
works again.  Still worries me as my last set of ssd disks got unstable
on one after less than 9 months of use and I'm probably about there with
these that are known to get cranky.  Smart reports them as ok, so I
wonder how bad ati taints the kernel space that it causes disk
controller/driver exceptions.

Moral of story: know when/how to repair whatever raid, as software and
hardware are seemingly still prone to exception from unlikely places.
Last time a disk died with md, I just mounted the secondary in an
enclosure, copied off data as pluggable, and copied to the new pair of
raid disks.  Hardware is never this easy, especially fakeraids.

-mb

I use software raid strictly on servers, which are headless (of course). I keep data on a server when practical, and run a daily rsync (offsite) backup of data on my workstation. Interesting to note the problems with video though. You might want to hop onto the raid list (gmane.linux.raid) and see what they think about it.

I don't know why anyone would run SSDs in a raid. RAID IS NOT A BACKUP. Of course if you're concerned about device failure and need to maintain continuity, raid is entirely appropriate.

One of the nice things about sw raid-1 is that as long as you have one good drive/partition, then you can recover. Just start the array in degraded mode, and you're good to go. I know that grub can access a single raid-1 partition w/out starting the array, which makes me wonder if you can simply mount one of the raid-1 partitions straight away w/out starting the array. I should try that so I know for sure if it's possible or not.

BL, *never* use fakeraid, and avoid raid-5 if possible. Disk space is no longer expensive enough to justify using raid-5.

--
-Eric 'shubes'

---------------------------------------------------
PLUG-discuss mailing list - PLUG-discuss@lists.plug.phoenix.az.us
To subscribe, unsubscribe, or to change your mail settings:
http://lists.PLUG.phoenix.az.us/mailman/listinfo/plug-discuss

Reply via email to