OK, that was the magic  incantation I  was looking for:
- changing the noprefetch option opened the floodgates to the L2ARC
- changing the max queue depth relived the wait time on the drives, although
I may undo this again in the benchmarking since these drives all have NCQ

I went from all four disks of the array at 100%, doing about 170 read
IOPS/25MB/s
to all four disks of the array at 0%, once hitting nealyr 500 IOPS/65MB/s
off the cache drive (@ only 50% load).
This bodes well for adding a second mirrored cache drive to push for the
1KIOPS.

Now I am ready to insert the mirror for the ZIL and the CACHE, and we will
be ready
for some production benchmarking.


 device    r/s    w/s   kr/s   kw/s wait actv  svc_t  %w  %b  us sy wt id
  sd0     170.0    0.4 7684.7    0.0  0.0 35.0  205.3   0 100  11  80  0 82
  sd1     168.4    0.4 7680.2    0.0  0.0 34.6  205.1   0 100
  sd2     172.0    0.4 7761.7    0.0  0.0 35.0  202.9   0 100
  sd4     170.0    0.4 7727.1    0.0  0.0 35.0  205.3   0 100
  sd5       1.6      2.6  182.4  104.8  0.0  0.5  117.8   0  31

                    extended device statistics
    r/s    w/s   kr/s   kw/s wait actv wsvc_t asvc_t  %w  %b device
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c0t0d0
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c0t0d1
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c0t0d2
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c0t0d3
  285.2    0.8 36236.2   14.4  0.0  0.5    0.0    1.8   1  37 c0t0d4


And, keep  in mind this was on less than $1000 of hardware.

Thanks,
Tracey


On Sat, Feb 13, 2010 at 9:22 AM, Richard Elling <richard.ell...@gmail.com>wrote:

> comment below...
>
> On Feb 12, 2010, at 2:25 PM, TMB wrote:
> > I have a similar question, I put together a cheapo RAID with four 1TB WD
> Black (7200) SATAs, in a 3TB RAIDZ1, and I added a 64GB OCZ Vertex SSD, with
> slice 0 (5GB) for ZIL and the rest of the SSD  for cache:
> > # zpool status dpool
> >  pool: dpool
> > state: ONLINE
> > scrub: none requested
> > config:
> >
> >        NAME        STATE     READ WRITE CKSUM
> >        dpool       ONLINE       0     0     0
> >          raidz1    ONLINE       0     0     0
> >            c0t0d0  ONLINE       0     0     0
> >            c0t0d1  ONLINE       0     0     0
> >            c0t0d2  ONLINE       0     0     0
> >            c0t0d3  ONLINE       0     0     0
> > [b]        logs
> >          c0t0d4s0  ONLINE       0     0     0[/b]
> > [b]        cache
> >          c0t0d4s1  ONLINE       0     0     0[/b]
> >        spares
> >          c0t0d6    AVAIL
> >          c0t0d7    AVAIL
> >
> >               capacity     operations    bandwidth
> > pool         used  avail   read  write   read  write
> > ----------  -----  -----  -----  -----  -----  -----
> > dpool       72.1G  3.55T    237     12  29.7M   597K
> >  raidz1    72.1G  3.55T    237      9  29.7M   469K
> >    c0t0d0      -      -    166      3  7.39M   157K
> >    c0t0d1      -      -    166      3  7.44M   157K
> >    c0t0d2      -      -    166      3  7.39M   157K
> >    c0t0d3      -      -    167      3  7.45M   157K
> >  c0t0d4s0    20K  4.97G      0      3      0   127K
> > cache           -      -      -      -      -      -
> >  c0t0d4s1  17.6G  36.4G      3      1   249K   119K
> > ----------  -----  -----  -----  -----  -----  -----
> > I just don't seem to be getting any bang for the buck I should be.  This
> was taken while rebuilding an Oracle index, all files stored in this pool.
>  The WD disks are at 100%, and nothing is coming from the cache.  The cache
> does have the entire DB cached (17.6G used), but hardly reads anything from
> it.  I also am not seeing the spike of data flowing into the ZIL either,
> although iostat show there is just write traffic hitting the SSD:
> >
> >                 extended device statistics                      cpu
> > device    r/s    w/s   kr/s   kw/s wait actv  svc_t  %w  %b  us sy wt id
> > sd0     170.0    0.4 7684.7    0.0  0.0 35.0  205.3   0 100  11  8  0 82
> > sd1     168.4    0.4 7680.2    0.0  0.0 34.6  205.1   0 100
> > sd2     172.0    0.4 7761.7    0.0  0.0 35.0  202.9   0 100
> > sd3       0.0      0.0      0.0    0.0  0.0  0.0    0.0   0   0
> > sd4     170.0    0.4 7727.1    0.0  0.0 35.0  205.3   0 100
> > [b]sd5       1.6      2.6  182.4  104.8  0.0  0.5  117.8   0  31 [/b]
>
> iostat has a "n" option, which is very useful for looking at device names
> :-)
>
> The SSD here is perfoming well.  The rest are clobbered. 205 millisecond
> response time will be agonizingly slow.
>
> By default, for this version of ZFS, up to 35 I/Os will be queued to the
> disk, which is why you see 35.0 in the "actv" column. The combination
> of actv=35 and svc_t>200 indicates that this is the place to start working.
> Begin by reducing zfs_vdev_max_pending from 35 to something like 1 to 4.
> This will reduce the concurrent load on the disks, thus reducing svc_t.
>
> http://www.solarisinternals.com/wiki/index.php/ZFS_Evil_Tuning_Guide#Device_I.2FO_Queue_Size_.28I.2FO_Concurrency.29
>
>  -- richard
>
> > Since this SSD is in a RAID array, and just presents as a regular disk
> LUN, is there a special incantation required to turn on the Turbo mode?
> >
> > Doesnt it seem that all this traffic should be maxing out the SSD? Reads
> from the cache, and writes to the ZIL? I have a seocnd identical SSD I
> wanted to add as a mirror, but it seems pointless if there's no zip to be
> had....
> >
> > help?
> >
> > Thanks,
> > Tracey
> > --
> > This message posted from opensolaris.org
> > _______________________________________________
> > zfs-discuss mailing list
> > zfs-discuss@opensolaris.org
> > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to