John, We just when through setting up our disk staging pool with TSM 5.1 on Solaris 8. We had 6x36G to work with.
The initial config used large files on the filesystem. This is incredibly slow. We then configured Disk Suite to mirror the disks and we configured TSM to use the raw devices e.g. /dev/md/rdsk/d0. Performance was significantly better. So TSM saw 3x36. From what we saw, it seemed the mirroring affected the performance more than the lack of spindles. As others have said, TSM seems to be pretty good at spreading the data around so it minimizes contention on the spool volumes. I never tested more, smaller mirrors. We then gave TSM the raw slices /dev/rdsk/c1t0d0s0 so he saw 6x36. This was the fastest by far. We were able to max the bandwidth (100M) for over an hour (38G in one hour). We lose the fault tolerance of mirrored disks, but we figure since it is only a staging area who cares? If a disk goes bad, we will lose the data backed up to it, but we can always back it up again. We felt the performance gains are well worth the redundancy hit. Though we have not tested pulling a disk from the staging pool and seeing what happens. I don't know your environment, but I would go with a single slice on all disks and tell TSM to use the raw device. If you really feel you need the redundancy I would just create 6 mirrors and use the raw mirrored device. From all of the benchmarking I've done it seems that once you get your setup decently tuned (don't tell TSM to use files for DB/LOG/DISKPOOLS) that the bottlenecks are either network capacity to your TSM server, or disk/cpu performance on the client (compression on). But I've only tested in one environment, ymmv. Hope this helps. scott Johnn D. Tan wrote: > I have 12 36-GB drives available for spool. > > Based on recommendations made to this list earlier this year, I went > with 12 mirrored disk spools of 16 GB each (keep in mind disk > overhead). > > As I understood it, the issue was you want many spools so that, as > Allen mentioned, you can have many threads for backups and even > migrations (assuming you have a good number of tape drives). > > However, you don't want so many spools per disk, otherwise there is > contention for head movement on the drive which would result in > poorer performance. > > johnn > > > >> => On Thu, 26 Sep 2002 08:54:01 -0400, Mahesh Tailor >> <[EMAIL PROTECTED]> said: >> >>> Hopefully this is a simple question: I have fourteen 36GB drives >>> that are >>> available for the diskpool and I was wondering whether it is better >>> to have >>> seven 5GB files or three 10GB files or one 35GB file or something >>> else? The >>> drives are mounted in two IBM-2014 Ultra-Wide SCSI disk drawers with >>> separate Ultra-Wide contollers. The other 14 drives are used for >>> DB, LOG, >>> and spare. >> >> >> You have a total of 28 spindles, 14 each on two busses, right? >> >> I'd suggest making a RAID-5 out of the fourteen free spindles, and >> then make >> the individual volumes "A reasonable size". What's a reasonable size? >> Uh... ;) >> >> I just did this with a drawer of 36G SSA, and I chose 10G volumes, >> because I >> have about a dozen (and growing) disk pools amongst which I need to >> divide >> things up. >> >> Even if you only have one or two disk pools, it's useful to have more >> than a >> few volumes per pool, because instantaneously only on thing can write >> to a >> volume at a time. So, for example, if you have 12 clients backing up, >> and one >> 70G disk volume, there is contention for the thread controlling that one >> volume. >> >> So calculate the size so that you'll have as many volumes as you feel >> like >> keeping track of, but not many more than that. >> >> >> - Allen S. Rout > > > -- Scott Walters Packet Pusher - "The world speaks IP" Mack Trucks, WHQ http://www.MackTrucks.com 2100 Mack Boulevard Ph: 610.709.3728 Allentown, PA 18103 Fx: 610.709.2809