Re: 3ware 9650 tips
On Mon, Jul 16, 2007 at 12:41:15PM +1000, David Chinner wrote: > On Fri, Jul 13, 2007 at 03:36:46PM -0400, Justin Piszcz wrote: > > On Fri, 13 Jul 2007, Jon Collette wrote: > > > > >Wouldn't Raid 6 be slower than Raid 5 because of the extra fault tolerance? > > > http://www.enterprisenetworksandservers.com/monthly/art.php?1754 - 20% > > >drop according to this article > > > > > >His 500GB WD drives are 7200RPM compared to the Raptors 10K. So his > > >numbers will be slower. > > >Justin what file system do you have running on the Raptors? I think thats > > >an interesting point made by Joshua. > > > > I use XFS: > > When it comes to bandwidth, there is good reason for that. > > > >>>Trying to stick with a supported config as much as possible, I need to > > >>>run ext3. As per usual, though, initial ext3 numbers are less than > > >>>impressive. Using bonnie++ to get a baseline, I get (after doing > > >>>'blockdev --setra 65536' on the device): > > >>>Write: 136MB/s > > >>>Read: 384MB/s > > >>> > > >>>Proving it's not the hardware, with XFS the numbers look like: > > >>>Write: 333MB/s > > >>>Read: 465MB/s > > >>> > > Those are pretty typical numbers. In my experience, ext3 is limited to about > 250MB/s buffered write speed. It's not disk limited, it's design limited. e.g. > on a disk subsystem where XFS was getting 4-5GB/s buffered write, ext3 was > doing > 250MB/s. > > http://oss.sgi.com/projects/xfs/papers/ols2006/ols-2006-paper.pdf > > If you've got any sort of serious disk array, ext3 is not the filesystem > to use To show what the difference is, I used blktrace and Chris Mason's seekwatcher script on a simple, single threaded dd command on a 12 disk dm RAID0 stripe: # dd if=/dev/zero of=/mnt/scratch/fred bs=1024k count=10k; sync http://oss.sgi.com/~dgc/writes/ext3_write.png http://oss.sgi.com/~dgc/writes/xfs_write.png You can see from the ext3 graph that it comes to a screeching halt every 5s (probably when pdflush runs) and at all other times the seek rate is >10,000 seeks/s. That's pretty bad for a brand new, empty filesystem and the only way it is sustained is the fact that the disks have their write caches turned on. ext4 will probably show better results, but I haven't got any of the tools installed to be able to test it The XFS pattern shows consistently an order of magnitude less seeks and consistent throughput above 600MB/s. To put the number of seeks in context, XFS is doing 512k I/Os at about 1200-1300 per second. The number of seeks? A bit above 10^3 per second or roughly 1 seek per I/O which is pretty much optimal. Cheers, Dave. -- Dave Chinner Principal Engineer SGI Australian Software Group - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: 3ware 9650 tips
On Monday 16 July 2007 14:22:25 David Chinner wrote: > You can see from the ext3 graph that it comes to a screeching halt > every 5s (probably when pdflush runs) and at all other times the > seek rate is >10,000 seeks/s. That's pretty bad for a brand new, > empty filesystem and the only way it is sustained is the fact that > the disks have their write caches turned on. ext4 will probably show > better results, but I haven't got any of the tools installed to be > able to test it I recently did some filesystem throuput tests, you may find it here http://www.pci.uni-heidelberg.de/tc/usr/bernd/downloads/lustre/performance/ ldiskfs is ext3+extents+mballoc+some-smaller-patches, so is almost ext4 (delayed allocation is still missing, but the clusterfs/lustre people didn't port it and I'm afraid of hard to detect filesystem corruptions if I include it myself). Write performance is still slower than with xfs and I'm really considering to try to use xfs in lustre. Cheers, Bernd -- Bernd Schubert Q-Leap Networks GmbH - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
MD RAID 5- EXT3-fs: group descriptors corrupted !
Hi folks, I am using RAID 5 ( Software MD RAID ) on my RAID controller card ( ata_piix). ( dmesg attached) - However, I am unable to mount my home partition over ext3. I get the errors EXT2-fs: group descriptors corrupted! EXT2-fs error (device md1): ext2_check_descriptors: Block bitmap for group 128 not in group (block 1630887232)! The output of lsraid is attached too - lsraid reports the devices to be just fine I tried fsck.ext3 -y /dev/md1 to no avail. Let me know if you need any more input Thanks in advance for all the help. Note i am using KUbuntu LTS 6.06, Kernel 2.6.15 Regards Ash [dev 9, 1] /dev/md1 5D39138F.4F5B7F2F.22B42D61.12EC18EB online [dev 8, 4] /dev/sda45D39138F.4F5B7F2F.22B42D61.12EC18EB good [dev 8, 19] /dev/sdb35D39138F.4F5B7F2F.22B42D61.12EC18EB good [dev 8, 35] /dev/sdc35D39138F.4F5B7F2F.22B42D61.12EC18EB good [dev 8, 51] /dev/sdd35D39138F.4F5B7F2F.22B42D61.12EC18EB spare [17179569.184000] Linux version 2.6.15-26-386 ([EMAIL PROTECTED]) (gcc version 4.0.3 (Ubuntu 4.0.3-1ubuntu5)) #1 PREEMPT Thu Aug 3 02:52:00 UTC 2006 [17179569.184000] BIOS-provided physical RAM map: [17179569.184000] BIOS-e820: - 0009fc00 (usable) [17179569.184000] BIOS-e820: 0009fc00 - 000a (reserved) [17179569.184000] BIOS-e820: 000e - 0010 (reserved) [17179569.184000] BIOS-e820: 0010 - 7fe76000 (usable) [17179569.184000] BIOS-e820: 7fe76000 - 7fee9000 (ACPI NVS) [17179569.184000] BIOS-e820: 7fee9000 - 7feed000 (usable) [17179569.184000] BIOS-e820: 7feed000 - 7feff000 (ACPI data) [17179569.184000] BIOS-e820: 7feff000 - 7ff0 (usable) [17179569.184000] 1151MB HIGHMEM available. [17179569.184000] 896MB LOWMEM available. [17179569.184000] found SMP MP-table at 000fe680 [17179569.184000] On node 0 totalpages: 524032 [17179569.184000] DMA zone: 4096 pages, LIFO batch:0 [17179569.184000] DMA32 zone: 0 pages, LIFO batch:0 [17179569.184000] Normal zone: 225280 pages, LIFO batch:31 [17179569.184000] HighMem zone: 294656 pages, LIFO batch:31 [17179569.184000] DMI 2.3 present. [17179569.184000] ACPI: RSDP (v000 INTEL ) @ 0x000fe020 [17179569.184000] ACPI: RSDT (v001 INTEL 04DT044 0x02a2 MSFT 0x0113) @ 0x7fefde48 [17179569.184000] ACPI: FADT (v001 INTEL 04DT044 0x02a2 MSFT 0x0113) @ 0x7fefcf10 [17179569.184000] ACPI: MADT (v001 INTEL 04DT044 0x02a2 MSFT 0x0113) @ 0x7fefce10 [17179569.184000] ACPI: WDDT (v001 INTEL 04DT044 0x02a2 MSFT 0x0113) @ 0x7fef7f90 [17179569.184000] ACPI: MCFG (v001 INTEL 04DT044 0x02a2 MSFT 0x0113) @ 0x7fef7f10 [17179569.184000] ACPI: ASF! (v032 INTEL 04DT044 0x02a2 MSFT 0x0113) @ 0x7fefcd10 [17179569.184000] ACPI: SSDT (v001 INTEL CpuPm 0x02a2 MSFT 0x0113) @ 0x7fefdc10 [17179569.184000] ACPI: SSDT (v001 INTEL Cpu0Ist 0x02a2 MSFT 0x0113) @ 0x7fefda10 [17179569.184000] ACPI: SSDT (v001 INTEL Cpu1Ist 0x02a2 MSFT 0x0113) @ 0x7fefd810 [17179569.184000] ACPI: SSDT (v001 INTEL Cpu2Ist 0x02a2 MSFT 0x0113) @ 0x7fefd610 [17179569.184000] ACPI: SSDT (v001 INTEL Cpu3Ist 0x02a2 MSFT 0x0113) @ 0x7fefd410 [17179569.184000] ACPI: DSDT (v001 INTEL 04DT044 0x02a2 MSFT 0x0113) @ 0x [17179569.184000] ACPI: PM-Timer IO Port: 0x408 [17179569.184000] ACPI: Local APIC address 0xfee0 [17179569.184000] ACPI: LAPIC (acpi_id[0x01] lapic_id[0x00] enabled) [17179569.184000] Processor #0 15:6 APIC version 20 [17179569.184000] ACPI: LAPIC (acpi_id[0x02] lapic_id[0x01] enabled) [17179569.184000] Processor #1 15:6 APIC version 20 [17179569.184000] WARNING: NR_CPUS limit of 1 reached. Processor ignored. [17179569.184000] ACPI: LAPIC (acpi_id[0x03] lapic_id[0x82] disabled) [17179569.184000] ACPI: LAPIC (acpi_id[0x04] lapic_id[0x83] disabled) [17179569.184000] ACPI: LAPIC_NMI (acpi_id[0x01] dfl dfl lint[0x1]) [17179569.184000] ACPI: LAPIC_NMI (acpi_id[0x02] dfl dfl lint[0x1]) [17179569.184000] ACPI: IOAPIC (id[0x05] address[0xfec0] gsi_base[0]) [17179569.184000] IOAPIC[0]: apic_id 5, version 32, address 0xfec0, GSI 0-23 [17179569.184000] ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) [17179569.184000] ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) [17179569.184000] ACPI: IRQ0 used by override. [17179569.184000] ACPI: IRQ2 used by override. [17179569.184000] ACPI: IRQ9 used by override. [17179569.184000] Enabling APIC mode: Flat. Using 1 I/O APICs [17179569.184000] Using ACPI (MADT) for SMP configuration information [17179569.184000] Allocating PCI resources starting at 8000 (gap: 7ff0:8010) [17179569.184000] Built 1 zonelists [17179569.184000] Kernel command line: root=/dev/md0 ro quiet splash [17179569.184000] mapped APIC to d000 (fee0) [17179569.184000] mapped IOAPIC to ff
RE: Software based SATA RAID-5 expandable arrays?
You will learn a lot by building your own system and will allow you to do more with it as far as other services if you want. However, again if you are still having problems with distro selection, configuration and commands, here is another NAS install solution I stumbled on. http://www.openfiler.com They appear to use a Fedora Distro, and remade it into their own. They also use the mdadm packages. I have not played with this, but If I had to chose, I would use this one since I have had more experience with mdadm as oppose to what the freenas is using. Their version of mdadm is not the very latest however. That won't effect you unless you want to be able to grow your RAID. You will need to update it. https://www.openfiler.com/community/forums/viewtopic.php?id=741 Oh, and they do support creating RAID6 arrays http://www.openfiler.com/screenshots/shots/RAID_Mgmt3.png Just giving you more options. Dan. - Original Message - From: Daniel Korstad Sent: Mon, 7/16/2007 7:48am To: Michael Subject: RE: Software based SATA RAID-5 expandable arrays? Something I ran across a year ago. http://www.freenas.org/index.php?option=com_versions&Itemid=51 I played with it for a day or so and it look impressive. The project is sill very much alive and they just released a new version a couple days ago. The caveat or reason I did not use this is that I use my Linux box for so many other things, (Web server, Asterisk (voip), Chillispot, VMware Server, Firewall, ... If you go this route, you will pretty much dedicate your box for just a NAS function. The project is an ISO OS you download and install. This greatly simplifies things but it ties you down a bit. After it is built, clients connect to it in server different options you can configure, CIFS (this is windows file sharing or samba), FTP, NFS, RSYNCD, SSHD, Unision, AFP. It also supports hard disk standby time, and advanced power management for your drives. However, if that is all you really want (a NAS) and you are having issues with other Linux distros... This is pretty simple to get one up and running with a NAS. Nice web interface for all the configuration. Other things to consider, I don't think it has RAID6. Or it did not last time I played with it a year ago. And I think the code is different than mdadm. So, you would be looking toward their forums for help if you had issues. Also, here is the manual for you.. http://www.freenas.org/downloads/docs/user-docs/FreeNAS-SUG.pdf Cheers, Dan. - Original Message - From: [EMAIL PROTECTED] on behalf of Daniel Korstad Sent: Fri, 7/13/2007 1:24pm To: big.green.jelly.bean Cc: davidsen ; linux-raid Subject: RE: Software based SATA RAID-5 expandable arrays? I can't speak for SuSe issues but I believe there is some confusion on the packages and command syntax. So hang on, we are going for a ride, step by step... Check and repair are not packages per say. You should have a package called echo. If you run this; echo 1 Should get a 1 echoed back at you. For example; [EMAIL PROTECTED] echo 1 1 Or anything else you want; [EMAIL PROTECTED] echo check check Now all we are doing with this is redirecting with the ">>" to another location, /sys/block/md0/md/sync_action The difference between a double >> and a single > is the >> will append it to the end and the single > will replace the contents of the file with the value. For example; I will create a file called foo; [EMAIL PROTECTED] tmp]# vi foo In this file I add two lines of text, foo, than I will write and quit :wq Now I will take a look at the file I just made with my vi editor... [EMAIL PROTECTED] tmp]# cat foo foo foo Great, now I run my echo command to send another value to it. First I use the double >> to just append; [EMAIL PROTECTED] tmp]# echo foo2 >> foo Now I take another look at the file; [EMAIL PROTECTED] tmp]# cat foo foo foo foo2 So, I have my first two text lines the third line "foo2" appended. Now I do this again but use just the single > to replace the file with a value. [EMAIL PROTECTED] tmp]# echo foo3 > foo Than I look at it again; [EMAIL PROTECTED] tmp]# cat foo foo3 Ahh, all the other lines are gone and now I just have foo3. So, > replaces and >> appends. How does this affect your /sys/block/md0/md/sync_action file? As it turns out, it does not matter. Think of the proc and sys (/proc and /sys) as psuedo file system is a real time, memory resident file system that tracks the processes running on your machine and the state of your system. So first lets go to /sys/block/ Than I will list its contents; [EMAIL PROTECTED] ~]# cd /sys/block/ [EMAIL PROTECTED] block]# ls dm-0 dm-3 hda md1 ram0 ram11 ram14 ram3 ram6 ram9 sdc sdf sdi dm-1 dm-4 hdc md2 ram1 ram12 ram15 ram4 ram7 sda sdd sdg dm-2 dm-5 md0 md3 ram10 ram13 ram2 ram5 ram8 sdb sde sdh This will be different for you since your system will have di
Re: Raid array is not automatically detected.
I do have the type set to 0xfd. Others have said that auto-assemble only works on RAID 0 and 1, but just as Justin mentioned, I too have another box with RAID5 that gets auto assembled by the kernel (also no initrd). I expected the same behavior when I built this array--again using mdadm instead of raidtools. Justin Piszcz wrote: On Sat, 14 Jul 2007, Bill Davidsen wrote: Bryan Christ wrote: My apologies if this is not the right place to ask this question. Hopefully it is. I created a RAID5 array with: mdadm --create /dev/md0 --level=5 --raid-devices=5 /dev/sda1 /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 mdadm -D /dev/md0 verifies the devices has a persistent super-block, but upon reboot, /dev/md0 does not get automatically assembled (an hence is not a installable/bootable device). I have created several raid1 arrays and one raid5 array this way and have never had this problem. In all fairness, this is the first time I have used mdadm for the job. Usually, I boot to something like SysRescueCD, used raidtools to create my array and then reboot with my Slackware install CD. Anyone know why this might be happening? Old type arrays are assembled due to having the proper partition type, 0xfd "Linux auto RAID" and are assembled by the kernel. All others are assembled by mdadm running out of initrd or similar, and failures there result from not having a proper config file in the initrd image. IIRC raidtools does set the array partitions to the auto-assemble partition type. Hope that points you in the right direction. Running "fdisk -l" as root will let you see all the partitions, types, etc, for everything on your system. I may be wrong, I thought auto-assemble only worked with type 0 or 1. -- bill davidsen <[EMAIL PROTECTED]> CTO TMR Associates, Inc Doing interesting things with small computers since 1979 - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html I use auto-assemble (in conjunction with Debian's own startup scripts) and for my root RAID1 device,swap and /boot, it is automatically taken care of by the kernel. For RAID5, it seems to work the same: [ 58.919378] RAID5 conf printout: [ 58.919418] --- rd:10 wd:10 [ 58.919457] disk 0, o:1, dev:sdc1 [ 58.919498] disk 1, o:1, dev:sdd1 [ 58.919539] disk 2, o:1, dev:sde1 [ 58.919579] disk 3, o:1, dev:sdf1 [ 58.919619] disk 4, o:1, dev:sdg1 [ 58.919659] disk 5, o:1, dev:sdh1 [ 58.919719] disk 6, o:1, dev:sdi1 [ 58.919759] disk 7, o:1, dev:sdj1 [ 58.919799] disk 8, o:1, dev:sdk1 [ 58.919839] disk 9, o:1, dev:sdl1 Justin. - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: 3ware 9650 tips
On Mon, 16 Jul 2007 at 12:41pm, David Chinner wrote If you've got any sort of serious disk array, ext3 is not the filesystem to use I do so wish that RedHat shared this view... -- Joshua Baker-LePain Department of Biomedical Engineering Duke University - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: Raid array is not automatically detected.
Bryan Christ wrote: I do have the type set to 0xfd. Others have said that auto-assemble only works on RAID 0 and 1, but just as Justin mentioned, I too have another box with RAID5 that gets auto assembled by the kernel (also no initrd). I expected the same behavior when I built this array--again using mdadm instead of raidtools. Any md arrays with partition type 0xfd using a 0.9 superblock should be auto-assembled by a standard kernel. If you want to boot from them you must ensure the kernel image is on a partition that the bootloader can read - ie RAID 0. This is nothing to do with auto-assembly. So some questions: * are the partitions 0xfd ? yes. * is the kernel standard? * are the superblocks version 0.9? (mdadm --examine /dev/component) David - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: 3ware 9650 tips
David Chinner wrote: > On Mon, Jul 16, 2007 at 12:41:15PM +1000, David Chinner wrote: >> On Fri, Jul 13, 2007 at 03:36:46PM -0400, Justin Piszcz wrote: ... >> If you've got any sort of serious disk array, ext3 is not the filesystem >> to use > > To show what the difference is, I used blktrace and Chris Mason's > seekwatcher script on a simple, single threaded dd command on > a 12 disk dm RAID0 stripe: > > # dd if=/dev/zero of=/mnt/scratch/fred bs=1024k count=10k; sync > > http://oss.sgi.com/~dgc/writes/ext3_write.png > http://oss.sgi.com/~dgc/writes/xfs_write.png Were those all with default mkfs & mount options? ext3 in writeback mode might be an interesting comparison too. -Eric - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: Software based SATA RAID-5 expandable arrays?
Due too the nature of the data I am storing RAID-6 is not really worth the extra safety and security, though it would be if I could get another 6 drives. Maybe then I can convert my RAID 5 into a RAID 6. As for openfiler, it is a great, simple package that provides all the features I need except that they dont include the latest kernel. That means my motherboard isnt supported. (frown). I have installed Fedora, after all the hastle of SuSe, and am currently setting that up so that it can be my main OS. It seems great, just some of the GUI based Admin tools are cryptic in their function. I have Mirrored my boot drive, which means I have to check to see if the second drive can be booted from. This is my todo list (though it does fail to mention SMART!), the times on the crontab have to be corrected. -- SAMBA http://www.redhatmagazine.com/2007/06/26/how-to-build-a-dirt-easy-home-nas-server-using-samba/ Repair http://www.issociate.de/board/post/391115/Observations_of_a_failing_disk.html http://www.issociate.de/board/post/443666/how_to_deal_with_continuously_getting_more_errors?.html Crontab (Weekly Repair Schedule) http://www.unixgeeks.org/security/newbie/unix/cron-1.html http://www.ss64.com/bash/crontab.html\ crontab -e 30 3 * * Mon echo check /sys/block/md3/md/sync_action 30 4 * * Mon echo check /sys/block/md0/md/sync_action 30 4 * * Mon echo check /sys/block/md1/md/sync_action 30 4 * * Mon echo check /sys/block/md2/md/sync_action Check Boot Info on Mirrored Drive After you go through the install and have a bootable OS that is running on mdadm RAID, I would test it to make sure grub was installed correctly to both the physical drives. If grub is not installed to both drives, and you lose one drive down the road and if that one was the one with grub, you will have a system that will not boot even though it has a second drive with a copy of all the files. If this were to happen, you can recover by booting with a bootable linux CD or recover disk and manually installing grub too. For example say you only had grub installed to hda and it failed, boot with a live linux cd and type (assuming /dev/hdd is the surviving second drive); grub device (hd0) /dev/hdd root (hd0,0) setup (hd0) quit System Report Email Mutt http://www.mutt.org/ http://linux.die.net/man/8/auditd.conf - Original Message From: Daniel Korstad <[EMAIL PROTECTED]> To: Michael <[EMAIL PROTECTED]> Cc: linux-raid@vger.kernel.org Sent: Monday, July 16, 2007 10:23:23 AM Subject: RE: Software based SATA RAID-5 expandable arrays? You will learn a lot by building your own system and will allow you to do more with it as far as other services if you want. However, again if you are still having problems with distro selection, configuration and commands, here is another NAS install solution I stumbled on. http://www.openfiler.com They appear to use a Fedora Distro, and remade it into their own. They also use the mdadm packages. I have not played with this, but If I had to chose, I would use this one since I have had more experience with mdadm as oppose to what the freenas is using. Their version of mdadm is not the very latest however. That won't effect you unless you want to be able to grow your RAID. You will need to update it. https://www.openfiler.com/community/forums/viewtopic.php?id=741 Oh, and they do support creating RAID6 arrays http://www.openfiler.com/screenshots/shots/RAID_Mgmt3.png Just giving you more options. Dan. - Original Message - From: Daniel Korstad Sent: Mon, 7/16/2007 7:48am To: Michael Subject: RE: Software based SATA RAID-5 expandable arrays? Something I ran across a year ago. http://www.freenas.org/index.php?option=com_versions&Itemid=51 I played with it for a day or so and it look impressive. The project is sill very much alive and they just released a new version a couple days ago. The caveat or reason I did not use this is that I use my Linux box for so many other things, (Web server, Asterisk (voip), Chillispot, VMware Server, Firewall, ... If you go this route, you will pretty much dedicate your box for just a NAS function. The project is an ISO OS you download and install. This greatly simplifies things but it ties you down a bit. After it is built, clients connect to it in server different options you can configure, CIFS (this is windows file sharing or samba), FTP, NFS, RSYNCD, SSHD, Unision, AFP. It also supports hard disk standby time, and advanced power management for your drives. However, if that is all you really want (a NAS) and you are having issues with other Linux distros... This is pretty simple to get one up and running with a NAS. Nice web interface for all the configuration. Other things to consider, I don't think it has RAID6. Or it did not last time I played with it a year ago. And I think the code is different than mdadm. So, you would be looking toward thei
Re: [Advocacy] Re: 3ware 9650 tips
On Mon, Jul 16, 2007 at 08:40:00PM +0300, Al Boldi wrote: > XFS surely rocks, but it's missing one critical component: data=ordered > And that's one component that's just too critical to overlook for an > enterprise environment that is built on data-integrity over performance. > > So that's the secret why people still use ext3, and XFS' reliance on external > hardware to ensure integrity is really misplaced. > > Now, maybe when we get the data=ordered onto the VFS level, then maybe XFS > may become viable for the enterprise, and ext3 may cease to be KING. Wow, thanks for bringing an advocacy thread onto linux-fsdevel. Just what we wanted. Do you have any insight into how to "get the data=ordered onto the VFS level"? Because to me, that sounds like pure nonsense. -- "Bill, look, we understand that you're interested in selling us this operating system, but compare it to ours. We can't possibly take such a retrograde step." - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: 3ware 9650 tips
On Mon, Jul 16, 2007 at 11:43:24AM -0400, Joshua Baker-LePain wrote: > On Mon, 16 Jul 2007 at 12:41pm, David Chinner wrote > > >If you've got any sort of serious disk array, ext3 is not the filesystem > >to use > > I do so wish that RedHat shared this view... So they support XFS in Fedora, but not in RHEL?? (I've been using Fedora...) - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
[RFC] VFS: data=ordered (was: [Advocacy] Re: 3ware 9650 tips)
Matthew Wilcox wrote: > On Mon, Jul 16, 2007 at 08:40:00PM +0300, Al Boldi wrote: > > XFS surely rocks, but it's missing one critical component: data=ordered > > And that's one component that's just too critical to overlook for an > > enterprise environment that is built on data-integrity over performance. > > > > So that's the secret why people still use ext3, and XFS' reliance on > > external hardware to ensure integrity is really misplaced. > > > > Now, maybe when we get the data=ordered onto the VFS level, then maybe > > XFS may become viable for the enterprise, and ext3 may cease to be KING. > > Wow, thanks for bringing an advocacy thread onto linux-fsdevel. Just what > we wanted. Do you have any insight into how to "get the data=ordered > onto the VFS level"? Because to me, that sounds like pure nonsense. Well, conceptually it sounds like a piece of cake, technically your guess is as good as mine. IIRC, akpm once mentioned something like this. But seriously, can you think of a technical reason why it shouldn't be possible to abstract data=ordered mode out into the VFS? Thanks! -- Al - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: [Advocacy] Re: 3ware 9650 tips
On Mon, 2007-07-16 at 11:48 -0600, Matthew Wilcox wrote: > Wow, thanks for bringing an advocacy thread onto linux-fsdevel. Just what > we wanted. Do you have any insight into how to "get the data=ordered > onto the VFS level"? Because to me, that sounds like pure nonsense. First off, I have no idea who decided to respond to my post and CC: linux-fsdevel on it. In retrospect, secondly, I should have not posted my post to linux-raid in the first place (is that list now mirrored to linux-fsdevel or something?). I was just sharing in my frustration of the lack of XFS support by Red Hat. So, lastly and in any case, my apologies to all, even if I did not proliferate it to linux-fsdevel, it was probably not ideal for me to post such to anything on vger.kernel.org (like linux-raid) in the first place. -- Bryan J. Smith Professional, Technical Annoyance mailto:[EMAIL PROTECTED] http://thebs413.blogspot.com Fission Power: An Inconvenient Solution - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: [RFC] VFS: data=ordered (was: [Advocacy] Re: 3ware 9650 tips)
On Mon, Jul 16, 2007 at 09:28:08PM +0300, Al Boldi wrote: > Well, conceptually it sounds like a piece of cake, technically your guess is > as good as mine. IIRC, akpm once mentioned something like this. How much have you looked at the VFS? There's nothing journalling-related in the VFS right now. ext3 and XFS share no common journalling code, nor do I think that would be possible, due to the very different concepts they have of journalling. Here's a good hint: $ find fs -type f |xargs grep -l journal_start fs/ext3/acl.c fs/ext3/inode.c fs/ext3/ioctl.c fs/ext3/namei.c fs/ext3/resize.c fs/ext3/super.c fs/ext3/xattr.c fs/ext4/acl.c fs/ext4/extents.c fs/ext4/inode.c fs/ext4/ioctl.c fs/ext4/namei.c fs/ext4/resize.c fs/ext4/super.c fs/ext4/xattr.c fs/jbd/journal.c fs/jbd/transaction.c fs/jbd2/journal.c fs/jbd2/transaction.c fs/ocfs2/journal.c fs/ocfs2/super.c JBD and JBD2 provide a journalling implementation that ext3, ext4 and ocfs2 use. Note that XFS doesn't, it has its own journalling code. If you want XFS to support data=ordered, talk to the XFS folks. Or start picking through XFS yourself, of course -- you do have the source code. -- "Bill, look, we understand that you're interested in selling us this operating system, but compare it to ours. We can't possibly take such a retrograde step." - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
RE: Software based SATA RAID-5 expandable arrays?
Don't forget the > or >> either one will do... crontab -e 30 3 * * Mon echo check > /sys/block/md3/md/sync_action 30 4 * * Mon echo check > /sys/block/md0/md/sync_action 30 4 * * Mon echo check > /sys/block/md1/md/sync_action 30 4 * * Mon echo check > /sys/block/md2/md/sync_action - Original Message - From: Michael Sent: Mon, 7/16/2007 12:34pm To: Daniel Korstad Cc: linux-raid@vger.kernel.org Subject: Re: Software based SATA RAID-5 expandable arrays? Due too the nature of the data I am storing RAID-6 is not really worth the extra safety and security, though it would be if I could get another 6 drives. Maybe then I can convert my RAID 5 into a RAID 6. As for openfiler, it is a great, simple package that provides all the features I need except that they dont include the latest kernel. That means my motherboard isnt supported. (frown). I have installed Fedora, after all the hastle of SuSe, and am currently setting that up so that it can be my main OS. It seems great, just some of the GUI based Admin tools are cryptic in their function. I have Mirrored my boot drive, which means I have to check to see if the second drive can be booted from. This is my todo list (though it does fail to mention SMART!), the times on the crontab have to be corrected. -- SAMBA http://www.redhatmagazine.com/2007/06/26/how-to-build-a-dirt-easy-home-nas-server-using-samba/ Repair http://www.issociate.de/board/post/391115/Observations_of_a_failing_disk.html http://www.issociate.de/board/post/443666/how_to_deal_with_continuously_getting_more_errors?.html Crontab (Weekly Repair Schedule) http://www.unixgeeks.org/security/newbie/unix/cron-1.html http://www.ss64.com/bash/crontab.html\ crontab -e 30 3 * * Mon echo check /sys/block/md3/md/sync_action 30 4 * * Mon echo check /sys/block/md0/md/sync_action 30 4 * * Mon echo check /sys/block/md1/md/sync_action 30 4 * * Mon echo check /sys/block/md2/md/sync_action Check Boot Info on Mirrored Drive After you go through the install and have a bootable OS that is running on mdadm RAID, I would test it to make sure grub was installed correctly to both the physical drives. If grub is not installed to both drives, and you lose one drive down the road and if that one was the one with grub, you will have a system that will not boot even though it has a second drive with a copy of all the files. If this were to happen, you can recover by booting with a bootable linux CD or recover disk and manually installing grub too. For example say you only had grub installed to hda and it failed, boot with a live linux cd and type (assuming /dev/hdd is the surviving second drive); grub device (hd0) /dev/hdd root (hd0,0) setup (hd0) quit System Report Email Mutt http://www.mutt.org/ http://linux.die.net/man/8/auditd.conf - Original Message From: Daniel Korstad <[EMAIL PROTECTED]> To: Michael <[EMAIL PROTECTED]> Cc: linux-raid@vger.kernel.org Sent: Monday, July 16, 2007 10:23:23 AM Subject: RE: Software based SATA RAID-5 expandable arrays? You will learn a lot by building your own system and will allow you to do more with it as far as other services if you want. However, again if you are still having problems with distro selection, configuration and commands, here is another NAS install solution I stumbled on. http://www.openfiler.com They appear to use a Fedora Distro, and remade it into their own. They also use the mdadm packages. I have not played with this, but If I had to chose, I would use this one since I have had more experience with mdadm as oppose to what the freenas is using. Their version of mdadm is not the very latest however. That won't effect you unless you want to be able to grow your RAID. You will need to update it. https://www.openfiler.com/community/forums/viewtopic.php?id=741 Oh, and they do support creating RAID6 arrays http://www.openfiler.com/screenshots/shots/RAID_Mgmt3.png Just giving you more options. Dan. - Original Message - From: Daniel Korstad Sent: Mon, 7/16/2007 7:48am To: Michael Subject: RE: Software based SATA RAID-5 expandable arrays? Something I ran across a year ago. http://www.freenas.org/index.php?option=com_versions&Itemid=51 I played with it for a day or so and it look impressive. The project is sill very much alive and they just released a new version a couple days ago. The caveat or reason I did not use this is that I use my Linux box for so many other things, (Web server, Asterisk (voip), Chillispot, VMware Server, Firewall, ... If you go this route, you will pretty much dedicate your box for just a NAS function. The project is an ISO OS you download and install. This greatly simplifies things but it ties you down a bit. After it is built, clients connect to it in server different options you can configure, CIFS (this is windows file sharing or samba), FTP, NFS, RSYNCD, SSHD, Unision, AFP. It also supports hard disk standb
Re: raid5:md3: read error corrected , followed by , Machine Check Exception: .
On Sat, Jul 14, 2007 at 07:32:38PM -0700, Mr. James W. Laferriere wrote: > >So your disk throws a fit > > Actually it's brand new . Infant mortallity ? I at least have a > cold spare available . So Yes I am replacing that puppy . I'll drop > it > into another system & give it the format command & see how much the user > bad block table grows . I'll bet I'll get a table full overflow on it . Manually keep both tables under supervision. I'd guess that if you send it a format, it will update the factory table (and move the user bad block table there). But most will retry writing to the bad sectors. And with a fully-fresh copy of the data, it will still be readable, and the blocks will be marked as ready-for-use, because that's better for performance > >And at some point at least 18 minutes after the raid incident you log > >CPU problems. > > I didn't notice the 18 Minute differance . Drats . I'm not sure how this happened, but the disk errror messages seem to have been logged by syslog, and the MCEs seem to have been copied from the console: They don't have the date attached? Roger. -- ** [EMAIL PROTECTED] ** http://www.BitWizard.nl/ ** +31-15-2600998 ** **Delftechpark 26 2628 XH Delft, The Netherlands. KVK: 27239233** *-- BitWizard writes Linux device drivers for any device you may have! --* Q: It doesn't work. A: Look buddy, doesn't work is an ambiguous statement. Does it sit on the couch all day? Is it unemployed? Please be specific! Define 'it' and what it isn't doing. - Adapted from lxrbot FAQ - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Software RAID across multiple SATA host controllers - OK?
Hi. I'm looking at setting up software RAID across 5 drives on an Intel motherboard that has a ICH7R 4-port SATA controller and also an additional 4 SATA ports on a Marvell controller. Is there anything I should be aware of when creating a software RAID5 array with drives hosted on more that one controller? Is this a bad idea, or no problem? Thanks -Greg - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: 3ware 9650 tips
On Mon, Jul 16, 2007 at 10:50:34AM -0500, Eric Sandeen wrote: > David Chinner wrote: > > On Mon, Jul 16, 2007 at 12:41:15PM +1000, David Chinner wrote: > >> On Fri, Jul 13, 2007 at 03:36:46PM -0400, Justin Piszcz wrote: > ... > >> If you've got any sort of serious disk array, ext3 is not the filesystem > >> to use > > > > To show what the difference is, I used blktrace and Chris Mason's > > seekwatcher script on a simple, single threaded dd command on > > a 12 disk dm RAID0 stripe: > > > > # dd if=/dev/zero of=/mnt/scratch/fred bs=1024k count=10k; sync > > > > http://oss.sgi.com/~dgc/writes/ext3_write.png > > http://oss.sgi.com/~dgc/writes/xfs_write.png > > Were those all with default mkfs & mount options? ext3 in writeback > mode might be an interesting comparison too. Defaults. i.e. # mkfs.ext3 /dev/mapper/dm0 # mkfs.xfs /dev/mapper/dm0 The mkfs.xfs picked up sunit/swidth correctly from the dm volume. Last time I checked, writeback made little difference to ext3 throughput; maybe 5-10% at most. I'll run it again later today... Cheers, Dave. -- Dave Chinner Principal Engineer SGI Australian Software Group - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html
Re: Software RAID across multiple SATA host controllers - OK?
On Mon, 16 Jul 2007, Greg Neumarke wrote: Hi. I'm looking at setting up software RAID across 5 drives on an Intel motherboard that has a ICH7R 4-port SATA controller and also an additional 4 SATA ports on a Marvell controller. Is there anything I should be aware of when creating a software RAID5 array with drives hosted on more that one controller? Is this a bad idea, or no problem? Thanks -Greg - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html I do it with 6 ports on 965 + 2 port sata controllers, no issues in my case. Justin. - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html