2008/1/28, Greg Freemyer <[EMAIL PROTECTED]>: > On Jan 28, 2008 11:25 AM, Ciro Iriarte <[EMAIL PROTECTED]> wrote: > > Hi, anybody has some notes about tuning md raid5, lvn and xfs?. I'm > > getting 20mb/s with dd and I think it can be improved. I'll add config > > parameters as soon as i get home. I'm using md raid5 on a motherboard > > with nvidia sata controller, 4x500gb samsung sata2 disks and lvm with > > OpenSUSE [EMAIL PROTECTED] > > > > Regards, > > Ciro > > -- > > I have not done any raid 5 perf. testing: 20 mb/sec seems pretty bad, > but not outrageous I suppose. I can get about 4-5GB/min from new sata > drives. So about 75 MB/sec from a single raw drive (ie. dd > if=/dev/zero of=/dev/sdb bs=4k) > > You don't say how your invoking dd. The default bs is only 512 bytes > I think and that is totally inefficient with the linux kernel. > > I typically use 4k which maps to what the kernel uses. ie. dd > if=/dev/zero of=big-file bs=4k count=1000 should give you a simple but > meaningful test.. > > I think the default stride is 64k per drive, so if your writing 3x 64K > at a time, you may get perfect alignment and miss the overhead of > having to recalculate the checksum all the time. > > As another data point, I would bump that up to 30x 64K and see if you > continue to get speed improvements. > > So tell us the write speed for > bs=512 > bs=4k > bs=192k > bs=1920k > > And the read speeds for the same. ie. dd if=big-file of=/dev/null bs=4k, > etc. > > I would expect the write speed to go up with each increase in bs, but > the read speed to be more or less constant. Then you need to figure > out what sort of real world block sizes your going to be using. Once > you have a bs, or collection of bs sizes that match your needs, then > you can start tuning your stack. > > Greg
Hi, posted the first mail from my cell phone, so couldn't add more info.... - I created the raid with chunk size= 256k. mainwks:~ # mdadm --misc --detail /dev/md2 /dev/md2: Version : 01.00.03 Creation Time : Sun Jan 27 20:08:48 2008 Raid Level : raid5 Array Size : 1465151232 (1397.28 GiB 1500.31 GB) Used Dev Size : 976767488 (465.76 GiB 500.10 GB) Raid Devices : 4 Total Devices : 4 Preferred Minor : 2 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Mon Jan 28 17:42:51 2008 State : active Active Devices : 4 Working Devices : 4 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 256K Name : 2 UUID : 65cb16de:d89af60e:6cac47da:88828cfe Events : 12 Number Major Minor RaidDevice State 0 8 33 0 active sync /dev/sdc1 1 8 49 1 active sync /dev/sdd1 2 8 65 2 active sync /dev/sde1 4 8 81 3 active sync /dev/sdf1 - Speed reported by hdparm: mainwks:~ # hdparm -tT /dev/sdc /dev/sdc: Timing cached reads: 1754 MB in 2.00 seconds = 877.60 MB/sec Timing buffered disk reads: 226 MB in 3.02 seconds = 74.76 MB/sec mainwks:~ # hdparm -tT /dev/md2 /dev/md2: Timing cached reads: 1250 MB in 2.00 seconds = 624.82 MB/sec Timing buffered disk reads: 620 MB in 3.01 seconds = 206.09 MB/sec - LVM: mainwks:~ # vgdisplay data Incorrect metadata area header checksum --- Volume group --- VG Name data System ID Format lvm2 Metadata Areas 1 Metadata Sequence No 5 VG Access read/write VG Status resizable MAX LV 0 Cur LV 2 Open LV 2 Max PV 0 Cur PV 1 Act PV 1 VG Size 1.36 TB PE Size 4.00 MB Total PE 357702 Alloc PE / Size 51200 / 200.00 GB Free PE / Size 306502 / 1.17 TB VG UUID KpUAeN-mPjO-2K8t-hiLX-FF0C-93R2-IP3aFI mainwks:~ # pvdisplay /dev/sdc1 Incorrect metadata area header checksum --- Physical volume --- PV Name /dev/md2 VG Name data PV Size 1.36 TB / not usable 3.75 MB Allocatable yes PE Size (KByte) 4096 Total PE 357702 Free PE 306502 Allocated PE 51200 PV UUID Axl2c0-RP95-WwO0-inHP-aJEF-6SYJ-Fqhnga - XFS: mainwks:~ # xfs_info /dev/data/test meta-data=/dev/mapper/data-test isize=256 agcount=16, agsize=1638400 blks = sectsz=512 attr=0 data = bsize=4096 blocks=26214400, imaxpct=25 = sunit=16 swidth=48 blks, unwritten=1 naming =version 2 bsize=4096 log =internal bsize=4096 blocks=16384, version=1 = sectsz=512 sunit=0 blks, lazy-count=0 realtime =none extsz=4096 blocks=0, rtextents=0 - The reported dd mainwks:~ # dd if=/dev/zero bs=1024k count=100 of=/mnt/custom/t3 100+0 records in 100+0 records out 104857600 bytes (105 MB) copied, 5.11596 s, 20.5 MB/s - New dd (seems to give better result) mainwks:~ # dd if=/dev/zero bs=1024k count=1000 of=/mnt/custom/t0 1000+0 records in 1000+0 records out 1048576000 bytes (1.0 GB) copied, 13.6218 s, 77.0 MB/s Ciro -- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]