Hi

I have troubles with my softwareraid. it keeps corrupting itself. i mean,
it's corrupt, i do "e2fsck /dev/md0", say yes to everything, filesystem
seems ok, work on it, filesystem is corrupt again...


the raid was fine during the last months running on rh7.3 and 8.0. now
i upgraded to rh9 and then the troubles started.

the upgrade went relatively smooth, just anaconda couldn't correctly figure
out my boot-drive to install grub. it suggested /dev/hda but it should be
/dev/sda. i don't know, if it touched hda (i did a nfs-install booting from
floppy, that's probably what got anaconda confused. it didn't see the higher
bootpriority of scsi over ide).
i then fixed grub manually to boot from sda.

during the upgrade my raid was disabled (deleted from fstab) because
anaconda couldn't mount it and aborted the upgrade.

after installation i installed all updates to rh9.

cat /proc/version
Linux version 2.4.20-18.8 ([EMAIL PROTECTED]) (gcc version 3.2 20020903 (Red Hat Linux 
8.0 3.2-7)) #1 Thu May 29 08:57:39 EDT 2003

the system is a 1GHz Athlon on a Asus A7V (i think). the system and bootdrive
is a scsi-disk connected to a ncr-controller. the raid is on 4 ide-disks
connected to the onboard via and promise controllers.

cat /etc/raidtab:
raiddev /dev/md0
    raid-level                5
    nr-raid-disks             4
    nr-spare-disks            0
    chunk-size                32
    persistent-superblock     1
    parity-algorithm          left-symmetric

    device                    /dev/hda1
    raid-disk                 0
    device                    /dev/hdc1
    raid-disk                 1
    device                    /dev/hde1
    raid-disk                 2
    device                    /dev/hdg1
    raid-disk                 3

when i brought the raid back up, it ran in degraded mode. /dev/hde1 was
kicked from the raid. so i cheched the drive with "hdparm -t -T" and
"badblocks /dev/hde", but it seemd fine. so i did "raidhotadd /dev/md0
/dev/hde1".

after waiting for reconstruction:

cat /proc/mdstat
Personalities : [raid5]
read_ahead 1024 sectors
md0 : active raid5 hdg1[3] hde1[2] hdc1[1] hda1[0]
      480238656 blocks level 5, 32k chunk, algorithm 2 [4/4] [UUUU]

unused devices: <none>

then i ran "e2fsck /dev/md0" which took a long time and tons of errors,
i had tons of stuff in lost+found and also most of the other files
were corrupt (could see bits from different files mixed together :(

of course i have no backup, but the data is not so valuable. it just takes
time to reencode all my cds again...

i tried to sort files out, but then the raid became corrupt agaiain, after
another fsck i tried again, but again the raid became corrupt.
that's where i am now.

i could start from scratch. ditch everything and recreate raid...

then i noticed this:
top -bn1
 14:00:02  up 1 day, 11:45,  4 users,  load average: 1.09, 1.11, 1.09
90 processes: 88 sleeping, 2 running, 0 zombie, 0 stopped
CPU states:   3.0% user   2.8% system  94.0% nice   0.0% iowait   0.0% idle
Mem:   514652k av,  506864k used,    7788k free,       0k shrd,  218388k buff
                    383288k actv,       0k in_d,   10192k in_c
Swap: 1049592k av,   22824k used, 1026768k free                  240800k cached

  PID USER     PRI  NI  SIZE  RSS SHARE STAT %CPU %MEM   TIME CPU COMMAND
[..]
    9 root     18446744073709551615 -20     0    0     0 SW<   0.0  0.0   0:00   0 
mdrecoveryd
[..]
 5694 root     18446744073709551615 -20     0    0     0 SW<   0.0  0.0   1:00   0 
raid5d
17801 root       9   0  1144  984   904 S     0.0  0.1   0:00   0 sshd
[..]

note the priority of mdrecoveryd and raid5d! (the sshd is just there to show
the normal priority of all the other processes).

what's wrong here?
does anybody successfully run a raid5 on kernel 2.4.20-18.8?
do you need more information?
how sould i proceed?

thanks for reading so far :)

bye
 dworz


-- 
redhat-list mailing list
unsubscribe mailto:[EMAIL PROTECTED]
https://www.redhat.com/mailman/listinfo/redhat-list

Reply via email to