So, assume your queue depth is 10, and you need a 1TB volume. You can present a 1TB LUN, and your whole SCSI queue for that volume is 10. If you present 2 500GB LUNs, your combined SCSI queue is 20 - one SCSI mailbox per LUN. If you present 10 100GB LUNs, you get a combined queue of 100. If you're doing small-block Oracle, with 95-98% cache read rates on the array (not unheard of), you can process more I/Os in-flight at any given point in time. You do make the valid point that making changes such as this can move the bottleneck from being storage-bound to being cpu-bound, and that's a not-uncommon result of this tuning on under-provisioned hardware. When I do this exercise with my customers, we probably run into that problem ~half the time.
As to stripe vs. concat specifically -- using a more artificial example, imagine you're putting together a 100GB volume. If you concatenate it as 2 50GB LUNs, you'll use one of the SCSI queues only, until you've filled up the first 50GB, then you'll start using both queues (mostly - this is an example). If you made the volume from, say, 10 10GB LUNs, you have a greater mathematical chance of any in-flight I/Os sorting across multiple queues, thus raising your overall I/O the application can request at any given time. Again, I want to stress that this does not mean this is *always* the right way to go, YMMV, please consult your local tuning professional or BOFH, etc. This is just an argument for why you'd ever _want_ to do it, using artificial examples. Most applications at most entities do not need this level of storage discipline. Those that do need it often stop falling over every night, by a simple storage change (or learn they need more box to meet SLAs, empirically). ZFS masks the gory details, but "ZFS" == "VxFS + VxVM", there are just different features in each product suite. Underneath the filesystem, it is still a RAID volume created on the host, that's managed by software - even if it's just one LUN. ZFS has features SF doesn't, and vice-versa, and we also have a CLI front-end that removes most of the complexity of the product for "simple" setups. My personal opinion is they fill different niches, ultimately, and neither is "better", just "better at some tasks". As to the Oracle issue, that's what the ODM API from Oracle solves, amongst other solutions . This article (http://blogs.sun.com/bobs/entry/one_i_o_two_i ) explains the benefits of concurrent, unbuffered I/O for Oracle better than I am able to. Cheers, jf As always, anything sent to this list are my own thoughts, and I speak only for myself. From: veritas-vx-boun...@mailman.eng.auburn.edu [mailto:veritas-vx-boun...@mailman.eng.auburn.edu] On Behalf Of Hudes, Dana Sent: Wednesday, January 06, 2010 11:10 AM To: William Havey Cc: veritas-vx@mailman.eng.auburn.edu Subject: Re: [Veritas-vx] Relayout volume from 2 to 4 columns So indeed this feature turns your $$$$$$$ Hitachi 9990V into a JBOD. Wow. I guess there are products that need this sort of thing. And enterprises where the people running the SAN array can't manage it properly so the server administrators have need of bypassing them. The other question of SCSI queues one per column as a benefit of striping is interesting. Doesn't this just generate extra disk i/o? The array is already doing RAID with 2 parity stripes. Now what? Yet this is effectively what ZFS does so there must be a performance gain. Hmm. Multiple SCSI queues actually might possibly make sense if you have a large number of CPUs (like the Sun 4v architecture esp. 5240 with 128 cores or 5440 with 256, or a 4+ board domain on the SunFire 25K which gives you 32 cores) all of which are running threads that do disk i/o. This benefit seems more practicaly in the ZFS approach where you have one volume-equivalent (the zpool is both disk group and VM volume in that it has storage layout) and many filesystems so you would likely have multiple processes doing independent disk i/o. In the VxVM one volume one filesstem model your e.g. Oracle table space is in one filesystem as one huge file (possibly other databases are files in other filesystems). Even if you have multiple listeners doing their thing ultimately there's one file they're working on...of course Oracle has row locking and other paralleization ....hmm. ________________________________ From: William Havey [mailto:bbha...@gmail.com] Sent: Wednesday, January 06, 2010 12:30 PM To: Hudes, Dana Cc: veritas-vx@mailman.eng.auburn.edu Subject: Re: [Veritas-vx] Relayout volume from 2 to 4 columns Yes, it certainly does. And that is why Symantec put the feature in the VM product; to use host-based software to construct and control storage from host to physical disk. This would help eliminate multi-vendor chaos in the storage aspects of the data center. On Wed, Jan 6, 2010 at 12:19 PM, Hudes, Dana <hud...@hra.nyc.gov> wrote: >the ISP feature of VM would allow you to drill down to individual spindles and place subdisks on each spindle. Individual spindles of the RAID group? Doesn't that defeat the purpose of the RAID group? Striping across LUNs gets ...interesting; we usually just use them concat. Of course that's with a real SAN array such as Hitachi 99x0 or Sun 61x0. I'm not sure I see the point of striping LUNs. If you are having performance problems from the array, fix the layout of the RAID group on the array: that's why you pay the big bucks to Hitachi for their hardware. I not sure I want to know about the load that could flatline a RAID-6 array of 6 15K RPM Fiber channel disks backed by a multigigabyte RAM cache. I have certainly seen bad storage layout on the host cause hot spots. That's when people make ridiculous numbers of small (gigabyte or so) volumes scattered all over the place -- another argument against the old way of doing things with databases and raw volumes (if you're going to use raw volumes at least use decent size ones not 2GB each ). While old (< 10) Solaris AIO did indeed suck dead bunnies thorugh a straw for performance, that's no longer a problem in Solaris 10 ZFS if you use it natively (using "branded" zones to run Solaris 8 and 9 puts the old AIO interface in front) nor would I expect it to be a problem with VxFS. ________________________________ From: veritas-vx-boun...@mailman.eng.auburn.edu [mailto:veritas-vx-boun...@mailman.eng.auburn.edu] On Behalf Of William Havey Sent: Wednesday, January 06, 2010 12:00 PM To: przemol...@poczta.fm Cc: veritas-vx@mailman.eng.auburn.edu Subject: Re: [Veritas-vx] Relayout volume from 2 to 4 columns VM views the two raid groups as single LUNs. It needn't be concerned with the layout of each raid group. To change from 2 columns to 4 columns use the relayout option to vxassist and also specify the two new LUNs on which to place the two new columns. That being said, the ISP feature of VM would allow you to drill down to individual spindles and place subdisks on each spindle. Bill On Wed, Jan 6, 2010 at 6:36 AM, <przemol...@poczta.fm> wrote: Hello, we are using VSF 5.0 MP3 on Solaris 10 attached to SAN-based hardware array. On this array we have created 2 raid groups and on each RG we have created a few LUNs: raid group: RG1 RG2 LUN1 LUN7 LUN2 LUN8 LUN3 LUN9 LUN4 LUN10 LUN5 LUN11 LUN6 LUN12 For performance reason some of our volumes are striped between the two raid groups (using two columns ncol=2) e.g.: pl <name> <vol> ENABLED ACTIVE 419256320 STRIPE 2/128 RW In this configuration IOs involves two raid groups. It seems that in the future in certain cases performance might be not as expected so we would like to add two additional LUNs (taken from two additional raid groups) and relayout the whole volume from 2-col to 4-cols e.g.: raid group: RG1 RG2 RG3 RG4 LUN1 LUN7 LUN13 LUN19 LUN2 LUN8 LUN14 LUN20 LUN3 LUN9 LUN15 LUN21 LUN4 LUN10 LUN16 LUN22 LUN5 LUN11 LUN17 LUN23 LUN6 LUN12 LUN18 LUN24 Is it possible to order relayout of existings volumes to spread it over all four RGs ? Can I point somehow that it should relayout using these particular LUNs ? Regards Przemyslaw Bak (przemol) -- http://przemol.blogspot.com/ ---------------------------------------------------------------------- Milosc, praca, pieniadze. Sprawdz swoj horoskop na dzis >> http://link.interia.pl/f2531 _______________________________________________ Veritas-vx maillist - Veritas-vx@mailman.eng.auburn.edu http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx _______________________________________________ Veritas-vx maillist - Veritas-vx@mailman.eng.auburn.edu http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx
_______________________________________________ Veritas-vx maillist - Veritas-vx@mailman.eng.auburn.edu http://mailman.eng.auburn.edu/mailman/listinfo/veritas-vx