2008/1/28, Greg Freemyer <[EMAIL PROTECTED]>:
> On Jan 28, 2008 11:25 AM, Ciro Iriarte <[EMAIL PROTECTED]> wrote:
> > Hi, anybody has some notes about tuning md raid5, lvn and xfs?. I'm
> > getting 20mb/s with dd and I think it can be improved. I'll add config
> > parameters as soon as i get home. I'm using md raid5 on a motherboard
> > with nvidia sata controller, 4x500gb samsung sata2 disks and lvm with
> > OpenSUSE [EMAIL PROTECTED]
> >
> > Regards,
> > Ciro
> > --
>
> I have not done any raid 5 perf. testing: 20 mb/sec seems pretty bad,
> but not outrageous I suppose.  I can get about 4-5GB/min from new sata
> drives.  So about 75 MB/sec from a single raw drive (ie. dd
> if=/dev/zero of=/dev/sdb bs=4k)
>
> You don't say how your invoking dd.  The default bs is only 512 bytes
> I think and that is totally inefficient with the linux kernel.
>
> I typically use 4k which maps to what the kernel uses.  ie. dd
> if=/dev/zero of=big-file bs=4k count=1000 should give you a simple but
> meaningful test..
>
> I think the default stride is 64k per drive, so if your writing 3x 64K
> at a time, you may get perfect alignment and miss the overhead of
> having to recalculate the checksum all the time.
>
> As another data point, I would bump that up to 30x 64K and see if you
> continue to get speed improvements.
>
> So tell us the write speed for
> bs=512
> bs=4k
> bs=192k
> bs=1920k
>
> And the read speeds for the same.  ie.  dd if=big-file of=/dev/null bs=4k, 
> etc.
>
> I would expect the write speed to go up with each increase in bs, but
> the read speed to be more or less constant.  Then you need to figure
> out what sort of real world block sizes your going to be using.  Once
> you have a bs, or collection of bs sizes that match your needs, then
> you can start tuning your stack.
>
> Greg

Hi, posted the first mail from my cell phone, so couldn't add more info....

- I created the raid with chunk size= 256k.

mainwks:~ # mdadm --misc --detail /dev/md2
/dev/md2:
        Version : 01.00.03
  Creation Time : Sun Jan 27 20:08:48 2008
     Raid Level : raid5
     Array Size : 1465151232 (1397.28 GiB 1500.31 GB)
  Used Dev Size : 976767488 (465.76 GiB 500.10 GB)
   Raid Devices : 4
  Total Devices : 4
Preferred Minor : 2
    Persistence : Superblock is persistent

  Intent Bitmap : Internal

    Update Time : Mon Jan 28 17:42:51 2008
          State : active
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 256K

           Name : 2
           UUID : 65cb16de:d89af60e:6cac47da:88828cfe
         Events : 12

    Number   Major   Minor   RaidDevice State
       0       8       33        0      active sync   /dev/sdc1
       1       8       49        1      active sync   /dev/sdd1
       2       8       65        2      active sync   /dev/sde1
       4       8       81        3      active sync   /dev/sdf1

- Speed reported by hdparm:

mainwks:~ # hdparm -tT /dev/sdc

/dev/sdc:
 Timing cached reads:   1754 MB in  2.00 seconds = 877.60 MB/sec
 Timing buffered disk reads:  226 MB in  3.02 seconds =  74.76 MB/sec
mainwks:~ # hdparm -tT /dev/md2

/dev/md2:
 Timing cached reads:   1250 MB in  2.00 seconds = 624.82 MB/sec
 Timing buffered disk reads:  620 MB in  3.01 seconds = 206.09 MB/sec

- LVM:

mainwks:~ # vgdisplay data
  Incorrect metadata area header checksum
  --- Volume group ---
  VG Name               data
  System ID
  Format                lvm2
  Metadata Areas        1
  Metadata Sequence No  5
  VG Access             read/write
  VG Status             resizable
  MAX LV                0
  Cur LV                2
  Open LV               2
  Max PV                0
  Cur PV                1
  Act PV                1
  VG Size               1.36 TB
  PE Size               4.00 MB
  Total PE              357702
  Alloc PE / Size       51200 / 200.00 GB
  Free  PE / Size       306502 / 1.17 TB
  VG UUID               KpUAeN-mPjO-2K8t-hiLX-FF0C-93R2-IP3aFI

mainwks:~ # pvdisplay /dev/sdc1
  Incorrect metadata area header checksum
  --- Physical volume ---
  PV Name               /dev/md2
  VG Name               data
  PV Size               1.36 TB / not usable 3.75 MB
  Allocatable           yes
  PE Size (KByte)       4096
  Total PE              357702
  Free PE               306502
  Allocated PE          51200
  PV UUID               Axl2c0-RP95-WwO0-inHP-aJEF-6SYJ-Fqhnga

- XFS:

mainwks:~ # xfs_info /dev/data/test
meta-data=/dev/mapper/data-test  isize=256    agcount=16, agsize=1638400 blks
         =                       sectsz=512   attr=0
data     =                       bsize=4096   blocks=26214400, imaxpct=25
         =                       sunit=16     swidth=48 blks, unwritten=1
naming   =version 2              bsize=4096
log      =internal               bsize=4096   blocks=16384, version=1
         =                       sectsz=512   sunit=0 blks, lazy-count=0
realtime =none                   extsz=4096   blocks=0, rtextents=0

- The reported dd
mainwks:~ # dd if=/dev/zero bs=1024k count=100 of=/mnt/custom/t3
100+0 records in
100+0 records out
104857600 bytes (105 MB) copied, 5.11596 s, 20.5 MB/s


- New dd (seems to give better result)
mainwks:~ # dd if=/dev/zero bs=1024k count=1000 of=/mnt/custom/t0
1000+0 records in
1000+0 records out
1048576000 bytes (1.0 GB) copied, 13.6218 s, 77.0 MB/s

Ciro
-- 
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to