Re: raid autodetect problem

2000-08-12 Thread Michael Robinton

You could probably make it work OK by using initrd and loading the module 
with the script, but compiling in support would be much more straightforward.

On Sat, 12 Aug 2000, Christian Bucher wrote:

> Hermann 'mrq1' Gausterer wrote:
> > 
> > if you have a raid-root-fs, then
> > i think the problem is that the kernel cannot load
> > the modules unless the rootfilessystem is mounted
> > and /lib/modules/ is available. but without
> > the module, a mount is not possible. :-o
> 
> Looks like raid-root-fs and /boot partition produces a similar situation.
> 
> 
> > have you tried with compiled in support ?
> 
> Of course it's working compiled into the kernel.
> 
> 
> 
> ciao
> 
> christian
> 
> -- 
> +---
> | Christian Bucher
> | phone: +43 1 5488640 mobil: +43 699 1 4208334
> | mailto:[EMAIL PROTECTED]
> +---
> 



Re: Failure autodetecting raid0 partitions

2000-07-16 Thread Michael Robinton

On Sat, 15 Jul 2000, Chris Mauritz wrote:

> > From [EMAIL PROTECTED] Sat Jul 15 19:29:44 2000
> > 
> > Edward Schernau wrote:
> > 
> > > Wow, an email CCed to Linus himself!  *faint*
> > 
> > Well do you know of another way to get a patch into the kernel ??
> 
> So if Linus gets hit by a bus (or a fast moving hari krishna), how
> are folks to get things into the kernel then?
> 
You send it to Alan Cox



Re: Monster machine and OpenLinux 2.3

2000-06-24 Thread Michael Robinton



On Sat, 24 Jun 2000, Bill Parker wrote:

>   I have a monster puter (Dual Pent-III 500, dual P/S (hot swap),
> 256MB ram, 9GB SCSI HD, AIC 7895 controller, 4x9GB Raid ARRAY (SA-1130),
> etc...I am thinking of installing linux on this beast, but would like to
> know if there is native RAID support for this machine...if you need more
> details, write back and I will do my best to supply info
> 
What kind of "native" did you have in mind. Linux software raid is pretty 
native to linux since 2.0x I run it on my dual Celron 500, 756mb ram, dual 
30gig segate 66ultras -- manages a nice 20meg xfer rate for read and 
write, not bad when the platter max is 26megs. I wouldn't exactly call 
this a monster, but the case is bigger than I like.



Re: LILO dies at LI

2000-06-12 Thread Michael Robinton

make sure and manually rdev your kernel

On Mon, 12 Jun 2000, Ryan Mack wrote:

> Running kernel 2.2.16-RAID (SMP) with LILO version 21 with attached
> raidtab (summary: mirrored /boot, striped /, both split across the same
> two drives). When I run lilo and reboot, lilo dies after printing LI To
> get my system to boot, I have to boot off of the RedHat CD-ROM (in rescue
> mode and mount my drives drives and run lilo with the redhat 2.2.14-12smp
> kernel. Then everything works groovy. Any ideas what's wrong? Thanks, Ryan
> 
> 



Re: Raid-1 boot problems

2000-05-30 Thread Michael Robinton

> 
>   I am currently having problems mounting my root file system that
> is set up with RAID-1. The initrd image I am using conatins all the raid
> tools, and successfully starts all the raid devices, but then I get the
> familiar message "Kernel panic: VFS: Unable to mount root fs on 00:00".
> The fstab on the initrd correctly shows /dev/md0 as /, and the
> "raidstart -a" command executed by the /linuxrc produces all the
> expected output. At the end, I still don't have a root filesystem.
> 
The usual problem is that you are booting with a kernel that has not had 
rdev done directly on the kernel -- lilo won't do it for you when you use 
initrd or boot from an image file.



linux-raid@vger.rutgers.edu

2000-05-14 Thread Michael Robinton

> We had some problems with raid-1 & lilo to allow booting
> from /dev/hda and /dev/hdc.
> 
See Boot+Root+Raid+Lilo at the Linux Documentation Project or

ftp.bizsystems.net/pub/raid/Boot+Root+Raid+LILO.html



Re: IDE Controllers

2000-05-09 Thread Michael Robinton



> 2.2.14 did not work for me.  I have a dual PIII with onboard UDMA33
> controller running RAID1.  Very stable.  When I just moved the
> two drives to a Promise Ultra66 the system became very unstable
> (uptimes in minutes). YMMV.

Hmmm I'm running 2.2.14 with latest Promise driver from their site.

Raid 5 running just fine.



Re: performance limitations of linux raid

2000-05-05 Thread Michael Robinton

> > > 
> > >   Not entirely, there is a fair bit more CPU overhead running an
> > > IDE bus than a proper SCSI one.
> > 
> > A "fair" bit on a 500mhz+ processor is really negligible.
> 
> 
>   Ehem, a fair bit on a 500Mhz CPU is ~ 30%.  I have watched a
> *single* UDMA66 drive (with read ahead, multiblock io, 32bit mode, and
> dma transfers enabled) on a 2.2.14 + IDE + RAID patched take over 30%
> of the CPU during disk activity.  The same system with a 4 x 28G RAID0
> set running would be < .1% idle during large copies.  An exactly
> configured system with UltraWide SCSI instead of IDE sits ~ 95% idle
> during the same ops.

Try turning on DMA



Re: performance limitations of linux raid

2000-05-04 Thread Michael Robinton

On Thu, 4 May 2000, Christopher E. Brown wrote:

> On Wed, 3 May 2000, Michael Robinton wrote:
> 
> > The primary limitation is probably the rotational speed of the disks and 
> > how fast you can rip data off the drives. For instance, the big IBM 
> > drives (20 - 40 gigs) have a limitation of about 27mbs for both the 7200 
> > and 10k rpm models. The Drives to come will have to make trade-offs 
> > between density and speed, as the technology's in the works have upper 
> > constraints on one or the other. So... given enough controllers (either 
> > scsii on disk or ide individual) the limit will be related to the 
> > bandwidth of the disk interface rather than the speed of the processor 
> > it's talking too.
> 
> 
>   Not entirely, there is a fair bit more CPU overhead running an
> IDE bus than a proper SCSI one.

A "fair" bit on a 500mhz+ processor is really negligible.
> 



Re: performance limitations of linux raid

2000-05-03 Thread Michael Robinton

The primary limitation is probably the rotational speed of the disks and 
how fast you can rip data off the drives. For instance, the big IBM 
drives (20 - 40 gigs) have a limitation of about 27mbs for both the 7200 
and 10k rpm models. The Drives to come will have to make trade-offs 
between density and speed, as the technologys in the works have upper 
constraints on one or the other. So... given enough controllers (either 
scsii on disk or ide individual) the limit will be related to the 
bandwidth of the disk interface rather than the speed of the processor 
it's talking too.

On Wed, 3 May 2000, Christopher E. Brown wrote:

> On Sun, 23 Apr 2000, Chris Mauritz wrote:
> 
> > > I wonder what the fastest speed any linux software raid has gotten, it
> > > would be great if the limitation was a hardware limitation i.e. cpu,
> > > (scsi/ide) interface speed, number of (scsi/ide) interfaces, drive
> > > speed. It would be interesting to see how close software raid could get
> > > to its hardware limitations.
> > 
> > RAID0 seemed to scale rather linearly.  I don't think there would be
> > much of a problem getting over 100mbits/sec on an array of 8-10 ultra2
> > wide drives.  I ultimately stopped fiddling with software RAID on my
> > production boxes as I needed something that would reliably do hot
> > swapping of dead drives.  So I've switched to using Mylex ExtremeRAID
> > 1100 cards instead (certainly not the card you want to use for low 
> > budget applications...heh).
> 
> 
>   Umm, I can get 13,000K/sec to/from ext2 from a *single*
> UltraWide Cheeta (best case, *long* reads, no seeks).  100Mbit is only
> 12,500K/sec.
> 
> 
>   A 4 drive UltraWide Cheeta array will top out an UltraWide bus
> at 40MByte/sec, over 3 times the max rate of a 100Mbit ethernet.
> 
> ---
> As folks might have suspected, not much survives except roaches, 
> and they don't carry large enough packets fast enough...
> --About the Internet and nuclear war.
> 
> 



Re: mkraid aborts without any useful info written to syslog

2000-05-02 Thread Michael Robinton



On Tue, 2 May 2000, Jeff Fookson wrote:

> I'm  having problems configuring a 2-disk  RAID-1 array on an  i686
> machine. The machine has Redhat 6.2 installed
> 
> 
>  cat /proc/mdstat
> Personalities : [1 linear] [2 raid0] [3 raid1] [4 raid5]
> read_ahead not set
> md0 : inactive
> md1 : inactive
> md2 : inactive
> md3 : inactive
> -

Old raid tools 
get a clean kernel from kernel.org -patch it with the latest raid patch 
as well as the promise patch u mentioned and you should be in good shape.

See:

ftp://ftp.bizsystems.net/pub/raid/
BootRootRaidLilominihowto

for details. You don't have to run root raid, but a synopsis is there as 
well as pointers to the current 0.90 raid tools Software HOWTO. Sorry I 
can't post it, I'm at home and dont'have the reference.



Re: performance limitations of linux raid

2000-04-23 Thread Michael Robinton


On Sun, 23 Apr 2000, bug1 wrote:
> Chris Mauritz wrote:
> > 
> > > Hi, im just wondering has anyone really explored the performance
> > > limitations of linux raid ?
> > 
> 
> Ive just managed to setup a 4 way ide raid0 that works.
> 
> The only way i can get it working is to use *two* drives per channel.
> I have to do this as i have concluded that i cannot use both my onboard
> hpt366 channels and my pci hpt366 channels together.
> 
> Ive done some superficial performance tests using dd, 55MB/s write
> 12MB/s read, interestingly i did get 42MB/s write using just a 2 way ide
> raid0, and got 55MB/s write with one drive per channel on four channels
> (i had no problem writing, just reading) so surprisingly i dont think
> the drive interface is my bottleneck.

> IDE will always beat SCSI hands down for price/performance, but scsi is
> clearly the winner if you want 4 or more drives on one machine, as ide
> just doesnt scale.

one of the ABIT boards comes with 4 onboard IDE controllers instead of 
the usual 2. As I recall, the kernel software will allow you to run up to 
8 ide pci channels, and the 2.4 kernel reportedly will do more.

Michael



Re: Tired of Spam

2000-04-09 Thread Michael Robinton

> > I'm tired of being spammed, and obviously the list admins are schmucks to 
> > leave the list open to spammers.
> > I have unsubscribed.
> 
> There are ways to deal with spammers, other than crippling the lists
> by closing them.
> 

Closing the list will not cripple it. I maintain a list with 2200+ 
members that grows about 50 - 100 every month. It is hardly crippled. 
What it allows us to do is zap the occasional spammer instantaneously.



Re: 2.2.12 & promise does not work

2000-04-08 Thread Michael Robinton

Check the promise site for new drivers

On Sat, 8 Apr 2000, octave klaba wrote:

> Hi,
> I have plus 3xpromise with 5x20Go ide to test raid5
> and I have some problems:
> - only 2 carts works !?
> - I run radhat6.1 and promise is not detected !?
> 
> pff :(((
> 
> have you any idea ?
> thanks
> 
> Octave
> 
> -- 
> Amicalement,
> Octave 
> > no swap allowed <
> 



Re: LILO error with a raid-1 /

2000-04-05 Thread Michael Robinton

On Wed, 5 Apr 2000, Sean Millichamp wrote:

> After that, I modified LILO to mount the root partition off of /dev/md0,
> ran lilo (it worked), and rebooted fine.  Less the 48 hours later I am
> trying to rerun lilo on the same machine and it won't run.  The only
> change I made were to delete the now useless boot targets for the old
> drives/partitions (that no longer exist).  This isn't critical right this
> instant but I am worried I won't be able to do routine things such as,
> say, upgrade my kernel without using a boot floppy (yuck).
> 
raid will boot quite nicely from a standard lilo if the config file is 
properly set up. I have several raid1 and raid5 (with raid one boot 
partions) running that use standard lilo on both scsi and ide systems. 
I'm at home at the moment so I will have to post a complete answer with 
examples config files + explainations tomorrow. Please be patient and 
I will get you the info.

Michael



Re: RAID-1 Non-destructive mirroring

2000-04-05 Thread Michael Robinton

It need not be destructive.

set up raid 1 mirroring with the original disk marked as failed, create 
the raid and run it. cp -a old disk / to /dev/mdx / set up lilo for 
raid1, make a boot diskett, boot from it and mount the raid1. fix the 
raidtab and raidhotadd the old disk to the array. reboot and you are in 
business.

Michael

On Wed, 5 Apr 2000 [EMAIL PROTECTED] wrote:

> 
> Is there any sign of RAID-1 that's non-destructive? This is making my
> life living hell right now. :)
> 
> I haven't been keeping up, so if there is a 2.2 patch or perhaps
> something in 2.3 that supports it, I would be most gracious (perhaps a method
> that isn't described in the howto?)
> 
> I would really like to be able to walk into a good portion of my
> clients and offer this, but the tedium of backup/partition/bootdisk/raid
> setup/restore is too much on a time-per-hour basis to offer them... I
> unfortunately have been having to offer "new technology" to long-standing
> UNIX clients who demand RAID-1 (which sickens me) to cover these needs since
> the price of hardware RAID controllers is even worse (and slower). 
> 
> Any help or kludges that can get me around this would be greatly
> appreciated.
> 
> -- 
> Erik Hollensbe
> I-Net Consultants
> [EMAIL PROTECTED]
> 



RE: Raid5 with two failed disks?

2000-04-02 Thread Michael Robinton

> Hmm, well, I'm certainly not positive why it wouldn't boot and I don't have
> the logs in front of me, but I do remember it saying that it couldn't mount
> /dev/md1 and therefore had a panic during boot. My solution was to specify
> the root device as /dev/sda1 instead of the configured /dev/md1 from the
> lilo prompt.

Hmm the only time I've seen this message has been when using initrd 
with an out of sync /dev/md or when the raidtab in the initrd was bad or 
missing. This was without autostart.

Michael


> 
> The disk is marked to auto raid start and marked as fd. And, it booted just
> fine until the "dumb" shutdown.
> 
> As for a rescue disk I'll put one together. Thanks for the advice.
> 
> --Rainer
> 
> 
> > -Original Message-
> > From: Michael Robinton [mailto:[EMAIL PROTECTED]]
> > Sent: Monday, April 03, 2000 8:50 AM
> > To: Rainer Mager
> > Cc: Jakob Ostergaard; [EMAIL PROTECTED]
> > Subject: RE: Raid5 with two failed disks?
> >
> > Whether or not the array is in sync should not make a difference to the
> > boot process. I have both raid1 and raid 5 systems that run root raid and
> > will boot quite nicely and rsync automatically after a "dumb" shutdown
> > that leaves them out of sync.
> >
> > Do you have your kernel built for auto raid start?? and partitions marked
> > "fd" ?
> >
> > You can reconstruct you existing array by booting with a kernel that
> > supports raid and with the raid tools on the rescue system. Do it all the
> > time.
> >
> > Michael
> >
> 



RE: Raid5 with two failed disks?

2000-04-02 Thread Michael Robinton

On Mon, 3 Apr 2000, Rainer Mager wrote:

>   I think my situation is the same as this "two failed disks" one but I
> haven't been following the thread carefully and I just want to double check.
> 
>   I have a mirrored RAID-1 setup between 2 disks with no spare disks.
> Inadvertantly the machine got powered down without a proper shutdown
> apparently causing the RAID to become unhappy. It would boot to the point
> where it needed to mount root and then would fail saying that it couldn't
> access /dev/md1 because the two RAID disks were out of sync.
>   Anyway, given this situation, how can I rebuild my array? Is all it takes
> is doing another mkraid (given the raidtab is identical to the real setup,
> etc)? If so, since I'm also booting off of raid, how do I do this for the
> boot partition? I can boot up using one of the individual disks (e.g.
> /dev/sda1) instead of the raid disk (/dev/md1), but if I do that will I be
> able to do a mkraid on an in-use partition? If not, how do I resolve this
> (boot from floppy?).
>   Finally, is there any way to automate this recovery process. That is, if
> the machine is improperly powered down again, can I have it automatically
> rebuild itself the next time it comes up?
> 
Whether or not the array is in sync should not make a difference to the 
boot process. I have both raid1 and raid 5 systems that run root raid and 
will boot quite nicely and rsync automatically after a "dumb" shutdown 
that leaves them out of sync.

Do you have your kernel built for auto raid start?? and partitions marked 
"fd" ?

You can reconstruct you existing array by booting with a kernel that 
supports raid and with the raid tools on the rescue system. Do it all the 
time.

Michael



suggested changes to raid tools

2000-04-02 Thread Michael Robinton

I'm in the process of upgrading another production system from old tools 
to new and noticed that you can't do a "raidstop" unless there is a valid 
raidtab. It seems to me that with a persistent superblock, this should 
not be necessary. The same applies to raidstart. Unless there is some 
reason to change the raid configuration -- i.e. remove or deactivate a 
piece, the superblock should provide all the necessary information. For 
raid to go mainstream, these minor things should be cleaned up. The docs 
seem to indicate that once the mkraid is done, the raidtab is really not 
needed. That is not true in the case with raidstop, raidstart, initrd 
starts and raid over raid (10, 51, etc...). All the info is there, the 
raidtab is redundant and will cause errors if not up todate or is missing.

Michael



RE: multiple partitions

2000-04-01 Thread Michael Robinton

On Sat, 1 Apr 2000 [EMAIL PROTECTED] wrote:

> On Fri, 31 Mar 2000, Michael wrote:
> 
> > hmmm. the remirroring code is not very smart... as I recall it 
> > does the remirroring in order .. i.e. md0, md1, etc... This would 
> > imply that if you have a power fail or other crash that causes both 
> > md's to be faulty, the system will not be able to swap until md0, 
> > then md1 is remirrored. The boot and swap md's should always be the 
> > first two raid partitions under this scenario as they are very small 
> > and will remirror quickly allowing swapping to proceed while the main 
> > array remirrors.
> 
> Isn't that exactly what we want to avoid? As I understand swapping on RAID
> while a resync is in progress is BAD.
> 
> But I agree this would be the way to go when this bug is fixed...
> 
You should read the whole thread. The purpose is to delay mounting the 
swap partition until the resync is complete on the raid/swap device, then 
proceed to mount it. When the resync is ordered, 0, 1, 2, this can result 
in very long delays if the swap device is not 0, since that would result 
in the main data partition being resynced first -- which takes a long 
time. Resyncing the swap partition of at most a few hundred megs takes 
a couple of minutes or seconds if you have a bodacious processor and very 
fast disks. This allows normal operation and swapping to proceed while 
the resync of the main array takes place in the background.



Re: raid parity and mmx processor

2000-03-19 Thread Michael Robinton


On Sun, 19 Mar 2000, Martin Eriksson wrote:

> - Original Message -
> From: "Michael Robinton" <[EMAIL PROTECTED]>
> To: <[EMAIL PROTECTED]>
> Sent: Sunday, March 19, 2000 9:54 AM
> Subject: raid parity and mmx processor
> 
> 
> > As I understand it, the kernel tests parity calc speed and decides to
> > use/not use mmx based on which is faster.
> >
> > Assuming an otherwise heavily loaded cpu, would it not be better to use
> > mmx even though slightly slower just to free up cpu cycles for other
> tasks.
> 
> If you are doing the calculations faster, it takes shorter time before the
> CPU is free to do other things. Fastest is fastest.
> 
Hmmm not really, if so, why would we have multi-processor systems.
mmx is a distinct processor apart from the fpp or main cpu. It is 
possible for all to be doing separate operations at the same time.
i.e. gzip, raid parity, network traffic


> >
> > If I don't have the right picture here, someone please provide an
> > explaination of where the cpu cycles go to raid.
> >
> > This reason for my question is the amount of "nice" cycles on one of my
> > machines that is fairly busy. Only raid5d is niced, and I'm running a
> > load factor over one consistently with only about a 20% cpu load. It
> > appears most of the cpu cycles in use go to raid5 -- presumably parity
> > calculations.
> 
> What kinda disks are you running on? SCSI? IDE? If IDE, do you have all nice
> things such as DMA transfers enabled?
> 
IDE 33 +dma

> >
> > Assuming I'm not too far off base, it would be nice to have a /proc
> > option or at least a compile option to force use of mmx for raid parity
> > and a way to run the parity calculation test manually to see the
> difference
> > between the two methods.
> 
> Maybe not for speed, but for bug tracing. Anyway, that would be good.
> 
> >



raid parity and mmx processor

2000-03-19 Thread Michael Robinton

As I understand it, the kernel tests parity calc speed and decides to 
use/not use mmx based on which is faster.

Assuming an otherwise heavily loaded cpu, would it not be better to use 
mmx even though slightly slower just to free up cpu cycles for other tasks.

If I don't have the right picture here, someone please provide an 
explaination of where the cpu cycles go to raid.

This reason for my question is the amount of "nice" cycles on one of my 
machines that is fairly busy. Only raid5d is niced, and I'm running a 
load factor over one consistently with only about a 20% cpu load. It 
appears most of the cpu cycles in use go to raid5 -- presumably parity 
calculations.

Assuming I'm not too far off base, it would be nice to have a /proc 
option or at least a compile option to force use of mmx for raid parity 
and a way to run the parity calculation test manually to see the difference 
between the two methods.

Michael



Re: Could one disk cause all this, or is it a bug in Linux-RAID?

2000-03-05 Thread Michael Robinton


On Sun, 5 Mar 2000, Johan Ekenberg wrote:

> > > About the crashes I reported yesterday:
> > > I received several mails suggesting cable problems,
> > > Power supply failure etc, so maybe I did not make
> > > my self clear enough: I've replaced all hardware
> > > except the disks. The disks have been checked
> >
> > How many disk controllers are involved here? If there is more than one
> > drive per controller it is "possible" for a single disk that has a scsi
> > bus problem to conceivably cause this -- scsi is supposed to be pretty
> > bullet proof (you did say scsi, don't have original post)
> 
> The Epox mb. has only one AIC7890 controller for U2W (LVD). All six drives
> are on this  controller.
> 
> Perhaps I have to replace all disks as well then...
> 
Hmm... scsii is pretty tough to mess up, but it is possible. Are there 
any error messages in any of the logs from the scsii bus? Still 
suspicious of a cable problem. I've had them where things at both ends 
work fine and the stuff in the middle throws errors.



Re: SV: Could one disk cause all this, or is it a bug in Linux-RAID?

2000-03-05 Thread Michael Robinton


On Sun, 5 Mar 2000, Johan Ekenberg wrote:

> About the crashes I reported yesterday:
> I received several mails suggesting cable problems,
> Power supply failure etc, so maybe I did not make
> my self clear enough: I've replaced all hardware
> except the disks. The disks have been checked

How many disk controllers are involved here? If there is more than one 
drive per controller it is "possible" for a single disk that has a scsi 
bus problem to conceivably cause this -- scsi is supposed to be pretty 
bullet proof (you did say scsi, don't have original post)



Re: question about reconstruction

2000-02-27 Thread Michael Robinton

> 
> > I plan to upgrade from old-style to new-style-raid, but on a 2.2.12 kernel.
> 
> Go to directly to 2.2.14, and use the patch announced on the list several
> times. It works correctly for me.
> 
I recently upgraded from 2.0.33 /0.42 tools to 2.2.14 /0.90 tools autostart
with no problem at all. I built a small raid first on swap partitions to 
proof the exact steps and when it went perfect, just did it on the 
production array. Worked great.



Re: RAID5 and 2.2.14

2000-01-24 Thread Michael Robinton


> Anyone successfully gotten raid 5 working with the new 2.2.14 kernel?
> I downloaded the 2.2.14 kernel source tree and installed it, then 
> downloaded Mingo's 2.2.14 raid patch.  The patch appeared to work fine on 
> the first few hunks, then failed miserably on the last 100 or so 

Yes, I converted an old 0.42 raid5 system last week and delivered a 
customer raid1 system, both with a virgin 2.2.14 kernel + mingos latest 
patch. No problems at all. The kernels were built on a clean 2.2.13 
(non-raid) system.

get your kernel from kernel.org
make sure you: patch -p1