Re: TAPEPOOL made up of mixed LTO2 LTO3 media
Bill, Can you post the output of q devc f=d? That may help us... You can keep your existing storage pools for your LTO2 tapes, and define new ones for your LTO3 tapes. That's what I did. In fact, I still purchase and use LTO2 cartridges, because of our DR contract which does not specify LTO3 drives, we send only LTO2 to our offsite vault. By local convention, I include the media type in the storage pool name. So, I have storage pools like WIN_LTO2, WIN_LTO3, and WIN_LTO2_COPY. The first two are primary pools, and the third is a copy pool (of course). I backup both of the primary pools to WIN_LTO2_COPY. I have four LTO3 drives and fourteen LTO2 drives in my STK library... TSM will use the LTO3 drives for LTO2 operations when they come up in the rotation... LTO3 operations will of course only use LTO3 drives. It is possible that all LTO3 drives can have LTO2 media loaded when an LTO3 operation starts, which causes that session/process to wait for mount points. Your original LTO (LTO2) storage pool will always have only LTO2 volumes, and your new LTO3 pool(s) will always have only LTO3 volumes. That's how storage pools work, all volumes must be of the same type. Which media gets used is determined by the destination in the copygroup... which points to a storage pool, which has a device class, which specifies the format, which determines the device type/recording density, etc... and the chosen media will be loaded into an appropriate drive. Hope this helps... _Robin Bill Dourado [EMAIL PROTECTED] TEC.COM To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: TAPEPOOL made up of mixed LTO2 LTO3 media 02/22/2007 05:13 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Hi, Thankyou to Robin and others for your useful comments. Is it possible to retain and continue using some of my existing tape storage pools associated with device class LTO ; and create new ones associated with device class LTO3 . ? Will the storage pools with LTO device class then use a mixture of LTO2 3 tapes ? And the new storage pools LTO3 tapes only ? Thanks Bill Robin Sharpe [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU 21/02/2007 19:37 Please respond to ADSM: Dist Stor Manager To: ADSM-L@VM.MARIST.EDU cc: Subject:Re: [ADSM-L] TAPEPOOL made up of mixed LTO2 LTO3 media If by tapepool, you mean storage pool, then I don't think so. A storage pool must be associated with one and only one device class. Each tape volume has an associated device class. Your existing LTO2 tapes probably use a device class with a format of ULTRIUM2C, which allows the highest capacity plus compression. When you introduce LTO3, you will probably want to use a device class with a format of ULTRIUM3C, to take advantage of LTO3's higher density. You will need a new storage pool (or pools) to hold your LTO3 tapes. It is (apparently) possible to update the device class with a new format... at least the help for the update devclass command implies it... but I don't know what would happen to an LTO2 tape with ACCESS=READWRITE if its devclass was updated to ULTRIUM3C format... and I wouldn't risk trying it. Whether or not your library can support mixed media is another question altogether. Some libraries can handle it seamlessly (like STK L700), and other need to be partitioned into logical libraries with each partition dedicated to the different media types. Robin Sharpe Berlex Labs Bill Dourado [EMAIL PROTECTED] TEC.COM To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU TAPEPOOL made up of mixed LTO2 LTO3 media 02/21/2007 06:57 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Hi, We are about to replace our all HP LTO-2 tape drives by LTO-3, in our NEO 4000 library. My question is, seeing the new drives can read and write both LTO-2 and LTO-3 media, can any given existing tapepool be made up of both LTO-2 and LTO-3 media ? My problem is that I don't have enough tape slots in the library and don't want to introduce
Re: TAPEPOOL made up of mixed LTO2 LTO3 media
If by tapepool, you mean storage pool, then I don't think so. A storage pool must be associated with one and only one device class. Each tape volume has an associated device class. Your existing LTO2 tapes probably use a device class with a format of ULTRIUM2C, which allows the highest capacity plus compression. When you introduce LTO3, you will probably want to use a device class with a format of ULTRIUM3C, to take advantage of LTO3's higher density. You will need a new storage pool (or pools) to hold your LTO3 tapes. It is (apparently) possible to update the device class with a new format... at least the help for the update devclass command implies it... but I don't know what would happen to an LTO2 tape with ACCESS=READWRITE if its devclass was updated to ULTRIUM3C format... and I wouldn't risk trying it. Whether or not your library can support mixed media is another question altogether. Some libraries can handle it seamlessly (like STK L700), and other need to be partitioned into logical libraries with each partition dedicated to the different media types. Robin Sharpe Berlex Labs Bill Dourado [EMAIL PROTECTED] TEC.COM To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU TAPEPOOL made up of mixed LTO2 LTO3 media 02/21/2007 06:57 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Hi, We are about to replace our all HP LTO-2 tape drives by LTO-3, in our NEO 4000 library. My question is, seeing the new drives can read and write both LTO-2 and LTO-3 media, can any given existing tapepool be made up of both LTO-2 and LTO-3 media ? My problem is that I don't have enough tape slots in the library and don't want to introduce a large number of LTO-3 tapes in one go. Also a few of our LTO-2 tapes are fairly new and it would be a shame to write them off. (TSM Server for Windows 5.3) Thanks Bill ** This electronic mail message, including any attachments, is a confidential communication exclusively between Babcock International Group PLC or its subsidiary company and the intended recipient(s) indicated as the addressee(s). It contains information which is private and may be proprietary or covered by legal professional privilege. If you receive this message in any form and you are not the intended recipient you must not review, use, disclose or disseminate it. We would be grateful if you could contact the sender upon receipt and in any event you should destroy this message without delay. Anything contained in this message that is not connected with the business of Babcock International Group PLC is neither endorsed by nor is the liability of this company. Babcock International Group PLC Company number 2342138 Registered in England 2 Cavendish Square London W1G 0PX Telephone: +44(0)20 7291 5000 Fax: +44(0)20 7291 5055 Website: www.babcock.co.uk **
Re: FW: Implementation of Active Only Pools in TSM V5.4
It sounds like ADP's are disk only (using FILE device type)... is this true? How are ADP's kept active-only... is it automatic, or do you have to run reclaims to remove inactive versions? Or do you just completely rebuild them periodically? Based on this description, I'm not sure this feature will be all that valuable... Robin Sharpe Berlex Labs Kelly Lipp [EMAIL PROTECTED] COM To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU FW: Implementation of Active Only Pools in TSM V5.4 02/04/2007 01:37 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Folks, I asked a question of TSM development regarding Active Data Pools (ADP) and received this excellent explanation. I think you will find it interesting. The one thing that escaped my attention in the documentation is the preference for restore choosing an ADP if one exists rather than a primary or standard copy pool. Thanks, Kelly J. Lipp VP Manufacturing CTO STORServer, Inc. 485-B Elkton Drive Colorado Springs, CO 80907 719-266-8777 [EMAIL PROTECTED] Active data pools (ADPs) represent a new type of storage pool, in addition to primary and copy pools. ADPs were designed to meet the following requirements. 1. Improved restore performance. ADPs can be used with FILE storage pools to allow rapid client restores. If available, files will preferentially be restored from a FILE ADP rather than a primary or copy pool. There is also some benefit when performing server operations such as GENERATE BACKUPSET. 2. Reduced resources for maintaining copy pools. In addition to being used for client restores, ADPs can also be used for storage pool or volume restore operations. Since an ADP only contains active data, the number of tapes is smaller than for a copy pool. Many customers want to be able to move tape volumes with only active data offsite for DR purposes and ADPs allow them to do so. By storing only active data, the number of offsite tapes is reduced as is the effort to manage those offsite tapes. 3. Reduced size of disk staging pools. Keeping only active versions of data in an ADP reduces the size of the disk pool as compared to keeping both active and inactive data in a copy pool. 4. Reduced data movement. This requirement stems from customers needing to stage data before performing client restore. Staging data required active and inactive data to be moved to disk before the store began. By keeping only active data, ADPs can eliminate the need for staging data to disk before restore. As an alternate design, TSM development also considered implementation of active-only primary pools. This implementation would not have supported restore of storage pools or volumes from the active-only primary volumes nor would it have facilitated movement of those active-only volumes to an offsite location. There were also significant technical issues involving migration (inactive data could not be removed from the active-only pool until it had been migrated) and with storage pool backup (backup from an active-only primary pool to a copy pool could mean that inactive files were not stored in the copy pool).
Re: Performance with move data and LTO3
Henrik, That's very interesting, and it looks like you're getting pretty good throughput overall. I wonder though, about tape 360024 on the Windows machine... it got much better throughput than any of the others on either machine. Do you think that was because of better compression on that volume? Regards, Robin Sharpe Berlex Labs Henrik Wahlstedt [EMAIL PROTECTED] To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: Performance with move data and LTO3 01/09/2007 10:26 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Hello, A late answer about LTO3 performance, but for the records. Origal post is below. I used 'audit vol volume_name skippartial=y fix=n' and 'move data volume_name reconstr=no'. Since the tape drives are not connected to a switch I used the information and timing from actlog when the tape is opened as an input volume and when the processing is done. Maybe not the best way but it will give me a clue about the performance. With IBM drives on Windows: select volume_name, est_capacity_mb, pct_utilized, pct_reclaim from volumes where volume_name='360023' VOLUME_NAME EST_CAPACITY_MB PCT_UTILIZED PCT_RECLAIM -- --- 360023 762938.0 34.2 0.0 select volume_name, est_capacity_mb, pct_utilized, pct_reclaim from volumes where volume_name='360024' VOLUME_NAME EST_CAPACITY_MB PCT_UTILIZED PCT_RECLAIM -- --- 360024 762938.0 49.6 0.0 select volume_name, est_capacity_mb, pct_utilized, pct_reclaim from volumes where volume_name='360125' VOLUME_NAME EST_CAPACITY_MB PCT_UTILIZED PCT_RECLAIM -- --- 360125 762938.0 45.6 0.0 360023 Audit: 196306 items / 33m04sec Move data: 196306 items / 273,902,675,481 bytes / 65m22sec ~69,84Mb/s 360024 Audit: 93202 items / 41m57sec Move data: 93202 items / 397,207,010,912 bytes / 53m29sec ~123,8Mb/s 360125 Audit: 21470 items / 71m02sec Move data: 21470 items / 369,229,158,737 bytes / 104m28sec ~58,9Mb/s With HP drives on Linux: select volume_name, est_capacity_mb, pct_utilized, pct_reclaim from volumes where volume_name='350075' VOLUME_NAME EST_CAPACITY_MB PCT_UTILIZED PCT_RECLAIM -- --- 350075 409600.0 41.6 0.0 select volume_name, est_capacity_mb, pct_utilized, pct_reclaim from volumes where volume_name='350204' VOLUME_NAME EST_CAPACITY_MB PCT_UTILIZED PCT_RECLAIM -- --- 350204 441710.3 68.9 31.1 select volume_name, est_capacity_mb, pct_utilized, pct_reclaim from volumes where volume_name='350257' VOLUME_NAME EST_CAPACITY_MB PCT_UTILIZED PCT_RECLAIM -- --- 350257 463674.8 49.5 50.6 350075 Audit: 705210 items / 38m30sec Move Data: 705210 items / 179,065,879,072 bytes / 36m48sec ~81,1Mb/s 350204 Audit: 36603 items / 71m35sec Move Data: 36603 items / 319,319,489,218 bytes / 75m11sec ~70,79mb/s 350257 Audit: 95345 items / 54m06sec Move Data: 95345 items / 240,739,702,948 bytes / 46m26sec
Re: Litigation! Wish
Now, that's cool -- getting competitive advantage from a product like TSM! I wonder if the API could be used to provide the much-sought-after restore preview function? My guess is no... I think I read somewhere that the API cannot access data backed up by the regular BA client. RS Steven Harris [EMAIL PROTECTED] IS.INFO To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: Litigation! Wish 12/26/2006 11:58 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Speaking of needles in haystacks, a one time colleague of mine was working for a company that analyzed oil seismic survey data on a large array of IBM clustered machines. He was a very smart cookie and understood the geophysics of it all (Hi Stephen if you are listening) The data came in on reels of tape and represented survey data from a surveyed line. A full survey consisted of a series of evenly spaced lines that mapped an area. He took this data and using the TSM API somehow stored it on 3590s (this was back in the old ADSM 3 days). The smart part was that oil companies could ask for an analysis of an area and give the coordinates that they wanted. His software would figure out which bits of data he needed from TSM mount the appropriate tapes and gather the data then feed it into the machines for analysis. This enabled his company to effectively leverage their investment in surveys and also provide the data faster to customers than anyone else could. At least that's what he told me :) The TSM API could be used to do a whole stack of this sort of storage work, but it is hampered by the lack of an API in something we can use, eg perl, python, or my current favourite ruby. I've taken a brief look at writing a library interface to ruby, but it is somewhat difficult - especially for an old COBOL programmer like me - why does everyone have to write in C anyway! Has anyone on the list done any work along these lines? Regards Steve Steven Harris AIX and TSM Admin Brisbane Australia
Re: Litigation! Wish
Well, what I meant was move data or move nodedata... but now that I think about it, those commands will have no effect on retention. They will only move the data to different volumes, maybe in different storage pools. A generate backupset would make a copy that has it's own retention criteria, but IMO backupsets are too hard to manage effectively... but then I haven't really used them that much. Also, backupsets will only contain active data, and so may be incomplete in a litigation context. I think the bottom line here, unfortunately, is that we're trying to make TSM fulfill a need it was not designed for. TSM is great for backing up a system and getting it back to a known operational state. It's also great for restoring a single file, set of files, directories, filesystems, etc. It's not too useful for finding a needle in a haystack, like we need all emails from John Doe to XYZ Corp regarding product X... there are archiving systems emerging that can do that kind of function. It would be nice if TSM could serve as the back end for such a system so you can minimize the back-end data store. I believe there are a couple that can do that. Of course, there's no free lunch... implementing an archive solution like that will cost significant bucks. -Robin Orin Rehorst [EMAIL PROTECTED] To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: Litigation! Wish 12/25/2006 11:40 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Thank you, Robin, for the confirmation of my suspicion. I understand special (second) backups with different parameters. Can you give me an example of or data movements within the server? TIA Orin Rehorst -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Robin Sharpe Sent: Thursday, December 21, 2006 8:55 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: Litigation! Wish Yes, this is a major shortcoming of TSM, especially in today's litigious business climate: the fact that data retention in TSM is tied to Management Class/Copy Group, and is the same throughout the storage hierarchy and across storage pool backups. The only way around it is to perform additional special backups and/or data movement within the server. Robin Sharpe Berlex Labs Orin Rehorst [EMAIL PROTECTED] To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Litigation! Wish 12/20/2006 09:57 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Added VTL to increase capacity. Dream wish: set Version Data Exists to 20 for vtlpool and to 2 for copypool. No way to accomplish that (there isn't a Santa, after all)? TIA Orin Rehorst
Re: Litigation! Wish
Yes, this is a major shortcoming of TSM, especially in today's litigious business climate: the fact that data retention in TSM is tied to Management Class/Copy Group, and is the same throughout the storage hierarchy and across storage pool backups. The only way around it is to perform additional special backups and/or data movement within the server. Robin Sharpe Berlex Labs Orin Rehorst [EMAIL PROTECTED] To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Litigation! Wish 12/20/2006 09:57 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Added VTL to increase capacity. Dream wish: set Version Data Exists to 20 for vtlpool and to 2 for copypool. No way to accomplish that (there isn't a Santa, after all)? TIA Orin Rehorst
Re: Compatiblity between LTO IBM and TSM drivers
My understanding is, you are supposed to use the IBM atdd drivers with IBM LTO drives. But you must use the TSM supplied tsmtape drivers with brand-X LTO drives. This may apply to other tape technologies as well, DLT for example. I have IBM LTO2 and LTO3 drives, and use the atdd drivers. At our DR site (HP), they have HP drives and I must use the tsmtape drivers there. Same is true for libraries: IBM libraries use the acdd driver, but brand-X (like my STK L700) use the tsmtape driver. Robin Sharpe Berlex Labs Schneider, John [EMAIL PROTECTED] Y.NET To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Compatiblity between LTO IBM and TSM drivers 12/12/2006 03:18 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Greetings, Quick compatibility question I haven't seen discussed. If it has and I missed it, I am sorry. My preference with IBM tape libraries and LTO drives on a TSM server is to use the IBMtape drivers, not the TSM drivers. In fact, I wonder why Tivoli even has a special driver for IBM's tape drives, when IBM provides a perfectly good one. I am about to upgrade a TSM server I just inherited from TSM 5.2.x to TSM 5.3.4 on a new box, and wanted to standardize on the IBMtape drivers from here on out. But I am also migrating a 3583 tape library with LTO1 drives and tapes that are already written with the TSM driver, and I am wondering if I am going to have a problem afterward being able to read the tapes written by the TSM driver. Can anyone tell me for certain if this is going to be a problem, or do I have to stay with the TSM driver? I don't really have anything against the TSM driver, I just want to make all my TSM servers the same configuation as much as I can. Best Regards, John D. Schneider Sr. System Administrator - Storage Sisters of Mercy Health System 3637 South Geyer Road St. Louis, MO. 63127 Email: [EMAIL PROTECTED] Office: 314-364-3150, Cell: 314-486-2359
Re: Trying to restore a clients volume (data) from one TSM server to another TSM server
If all you need to do is a one time restore, I would just register the target client node (DOCDRB) on TSM server A, perform the restore, them remove the definition of DOCDRB. Regards, Robin Sharpe Berlex Labs Timothy Hughes [EMAIL PROTECTED] IT.STATE.NJ.USTo Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Trying to restore a clients volume (data) from one TSM server to another TSM server 12/11/2006 10:10 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU A client tried to restore a volume's data ( NVOL1:) from DOCUSRA to a volume on DOCDRB but can't. DOCUSRA is a node registered on TSM server A and DOCDRSB is registered on TSM server B does something needs to be configured for this restore to work? I understand I can restore data from one node to another on the same TSM server but what would the syntax be to restore data to another node on another TSM server? for DR purposes? Tim ANS1096S Either the node does not exist on the server or there is no active policy set for the node. Explanation: This error occurs when you try to access another node's data. Either the node is not registered with the TSM server, or there is no active policy set for the node. Thanks in advance for any help or suggestings. TSM 5.3.4.0 Novell client 5.3.4.0 AIX 5.3 RS/6000
Re: Trying to restore a clients volume (data) from one TSM server to another TSM server
OK, if this is a DR scenario, it's a little more complicated. In fact, a lot more complicated. Are both TSM servers at the same location? I wouldn't expect they are if this is for DR. In a disaster, you need to plan for TSM A being unavailable, so you need a way to get your nodes' data to TSM B in a basic DR plan this is done by doing a TSM database restore... documented in detail in the Admin Guide. Once that is done, you can restore your clients at the DR site. It sounds, though, like you have two running TSM servers, A and B. Are you trying to have them be a DR TSM server for each other? If your hardware is robust enough, you could plan to restore TSM A on the TSM B hardware as a second instance so TSM A and B would both run on the same physical server (from the disaster event until the original site is rebuilt). Many of us run this way because of capacity issues... I have five TSM instances on one HP-UX server, sharing a SCSI library. I guess the bottom line -- the bad news -- is that there is not a way to do exactly what you are describing (AFAIK). TSM B knows nothing about TSM A, or it's nodes or storage pools or volumes. Therefore you can't restore a node which normally belongs to TSM B with data from a corresponding node on TSM A. You can register the target node on TSM A and restore directly, but you cannot count on TSM A being available in the event of a disaster. (I'm thinking while I type) How about this: TSM A at site A, TSM B at site B. All nodes are defined to both TSM A and B. TSM A at site A backs up all of the nodes at site B over the network. (big assumption here that your network can support the traffic) TSM B backs up nodes at site A. Disaster occurs, site A burns to the ground. Restore site A servers to their DR counterparts at site B using their backups on TSM B. Start backing up site B nodes to TSM B. But, you have lost all of site B nodes' existing backups in the fire unless you send copy pool volumes offsite. You still need to do a TSM DR database restore to get that history back. Rudimentary at best... Timothy Hughes [EMAIL PROTECTED] IT.STATE.NJ.USTo Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: Trying to restore a clients volume (data) from one TSM server to another TSM server 12/11/2006 01:41 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Robin here's a further explanation. Is there a way to access the TSM backup server for the node (s) that I want the data restored to? There are six 6 nodes the production set on TSM server A and the DR set on TSM server B. The client would like to be able to restore data from one node to it's matching node on TSM server B. Is there a parameter that has to be configured? Thanks again Robin Sharpe wrote: If all you need to do is a one time restore, I would just register the target client node (DOCDRB) on TSM server A, perform the restore, them remove the definition of DOCDRB. Regards, Robin Sharpe Berlex Labs Timothy Hughes [EMAIL PROTECTED] IT.STATE.NJ.USTo Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Trying to restore a clients volume (data) from one TSM server to another TSM server 12/11/2006 10:10 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU A client tried to restore a volume's data ( NVOL1:) from DOCUSRA to a volume on DOCDRB but can't. DOCUSRA is a node registered on TSM server A and DOCDRSB is registered on TSM server B does something needs to be configured for this restore to work? I understand I can restore data from one node to another on the same TSM server but what would the syntax be to restore data to another node on another TSM server? for DR purposes? Tim ANS1096S Either the node does not exist on the server or there is no active policy set for the node. Explanation: This error occurs when you try to access another node's data. Either the node is not registered with the TSM server, or there is no active policy set for the node. Thanks in advance for any
Re: IP address
Avy, If you need the name as configured on the server, just the q session command from any BA client with that IP address in its dsm.sys (or dsm.opt) file, like the following example from one of my Unix clients. The server name in this case is TSM_WAYNE_DR. wau004:/home/root dsmc q se Tivoli Storage Manager Command Line Backup/Archive Client Interface - Version 5, Release 1, Level 1.0 (C) Copyright IBM Corporation 1990, 2002 All Rights Reserved. Node Name: WAU004 Session established with server TSM_WAYNE_DR: HP-UX Server Version 5, Release 3, Level 2.0 Server date/time: 11/27/06 08:51:08 Last access: 11/27/06 08:50:04 TSM Server Connection Information Server Name.: TSM_WAYNE_DR Server Type.: HP-UX Server Version..: Ver. 5, Rel. 3, Lev. 2.0 Last Access Date: 11/27/06 08:50:04 Delete Backup Files.: No Delete Archive Files: Yes Node Name...: WAU004 User Name...: root wau004:/home/root _RS Avy Wong [EMAIL PROTECTED] .COM To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU IP address 11/24/2006 03:52 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU hello, Can you tell me if there is a way to find out what is the server name when all I have is an ip address? is there a query to do that? Thanks. Avy Wong Business Continuity Administrator Mohegan Sun 1 Mohegan Sun Blvd Uncasville, CT 06382 ext 28164 ** The information contained in this message may be privileged and confidential and protected from disclosure. If the reader of this message is not the intended recipient, or an employee or agent responsible for delivering this message to the intended recipient, you are hereby notified that any dissemination, distribution, or copy of this communication is strictly prohibited. If you have received this communication in error, please notify us immediately by replying to the message and deleting it from your computer. **
Re: Strange Reclamation Problem
Andrew, Are you absolutely SURE that some data did not get migrated to tape? It looks like that's what happened... maybe do some q content commands on the disk pool, or select from volumeusage for one of the nodes that did not complete... Robin Sharpe Berlex Labs ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 09/18/2006 04:07:57 PM: I am running TSM 5.3.2.3 on an AIX (5.2.5) platform. We recently finished migrating our onsite pool to disk only, and out offsite pool to directly attached 3592 drives in a 3584 silo. We started getting multiple ANR1163I messages. I started investigating this, and found an odd behaviour. If I entered a move volume command for one of the volumes, it would move some data, then end, with no errors. It appears to be moving the data of one node off the cart, then ending the process. These are volumes that were previously not collocated. If I update the volume to read-write (I forgot to mention I make them offsite so the TSM will read from the disk pool), and do the move volume tape-to-tape, it works fine. It finishes the whole tape. Any ideas? I am going to open a PMR with IBM, but thought I would ask here first in case I missed something incredibly obvious. Thanks. ANR1040I Space reclamation started for volume 110021, storage pool COPY3592 (process number 1555). ANR1040I Space reclamation started for volume 110482, storage pool COPY3592 (process number 1555). ANR1040I Space reclamation started for volume 110436, storage pool COPY3592 (process number 1555). ANR1040I Space reclamation started for volume 110268, storage pool COPY3592 (process number 1555). ANR1040I Space reclamation started for volume 111027, storage pool COPY3592 (process number 1555). ANR8468I 3592 volume 110192 dismounted from drive RMT5 (/dev/rmt5) in library 3584LIB. ANR0409I Session 349997 ended for server TSMLIBM (AIX-RS/6000). ANR1163W Offsite volume 110212 still contains files which could not be moved. ANR1163W Offsite volume 110499 still contains files which could not be moved. ANR1163W Offsite volume 110163 still contains files which could not be moved. ANR1163W Offsite volume 110021 still contains files which could not be moved. ANR1163W Offsite volume 110482 still contains files which could not be moved. ANR1163W Offsite volume 110436 still contains files which could not be moved. ANR1163W Offsite volume 110268 still contains files which could not be moved. ANR1163W Offsite volume 111027 still contains files which could not be moved. ANR4932I Reclamation process 1555 ended for storage pool COPY3592. ANR0986I Process 1555 for SPACE RECLAMATION running in the BACKGROUND processed 6335 items for a total of 1,768,794,536 bytes with a completion state of SUCCESS at 10:13:52. Andy Carlson --- Gamecube:$150,PSO:$50,Broadband Adapter: $35, Hunters License: $8.95 /month, The feeling of seeing the red box with the item you want in it:Priceless.
Re: MAXNUMMP
Well, I'd say that seems to be what is causing your intermittent failures then. Unfortunately, there is no magic bullet approach to fix this situation -- it requires cooperation of all the admins involved (TSM, DBA, Unix, applications), and the TSM admin has the responsibility to educate all parties about the interactions. For example, the DBAs must be made aware that if they set their parallelism (is that the right term?) too high (higher than MAXNUMMP), some channels will not be able to work, and RMAN jobs will fail. Do NOT set MAXNUMMP higher than the number of installed drives that will almost guarantee failures. If you set it equal to the number of installed drives, then all of those drives must be available for that node when it wants them, or there will be failures. It requires coordinated scheduling. The approach I would take is to set MAXNUMMP only as high as that client needs to get its backup done in the time allotted if a particular node MUST backup 100GB in ten minutes (as an absurd example), it will need several drives... but if it has four hours to complete its backup, then one drive is plenty. Another approach, if you have it available, is to use an external scheduler (such as Control-M) rather than the TSM scheduler. Most enterprise class schedulers can manage the drives as a resource pool, and will only start a backup that needs four drives if four drives are actually available. This is a labor-intensive approach (initially), and it still is not fool-proof. The approach we use, is that ALL backups go to a disk pool initially. Nothing goes directly to tape. Disk pools do not use the MAXNUMMP value, so you can run as many channels and sessions as your hardware/OS/TSM can handle. This also eliminates the shoe-shining problem with streaming tape drives such as LTO and DLT (or at least postpones it). However, this does introduce another problem, at least for TDP Oracle clients... if the disk pool fills up, TDPO will not go to the next pool in the hierarchy like the BA client does... it will fail. In practice this means keeping the migration threshold low enough on those disk pools so that there will always be enough space available for TDPO. Again, this requires some careful analysis and cooperation of the TSM admin and the DBA team. Sorry for the long-winded response, I was on a roll! Hope it helps... Robin Sharpe Berlex Labs ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 09/14/2006 04:29:48 PM: Is it possible that there are two or more sessions running for the TDP client simultaneously? Yes, absolutely. The Oracle DBAs have observed that this happens most frequently when there's media wait, in which case multiple log backups could be running simultaneously. How do others handle this? Does it make sense to set MAXNUMMP to the number of drives...or even higher? I remember being very unhappy when I set the value of MAXNUMMP to the number of drives, but I can't remember what happened. Maybe I was running into some other problem. Thanks again for all the ideas, anker
Re: MAXNUMMP
Anker, Is it possible that there are two or more sessions running for the TDP client simultaneously? The MAXNUMMP value is total for the node, so if it is '1', and you start two sessions (e.g. two RMAN scripts), one of them will not have a mount point available... Regards, Robin Sharpe Berlex Labs ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 09/14/2006 11:53:38 AM: ORA-19511: Error received from media manager layer, error text: ANS0326E (RC41) This node has exceeded its maximum number of mount points. Actually, I don't understand why these backups are calling for multiple drives anyway. They're routine Oracle log backups that happen many times a day and run successfully 99.9% of the time. Similarly, the Windows backups are routine nightly backups. I'm not trying to stream to multiple tapes. Why shouldn't MAXNUMMP=1 do the job?
LTO-2 and LTO-3 switch grouping
Hi All, We have the good fortune to be adding four LTO-3 drives to our STK L700 library. It currently has 14 LTO-2 drives, attached to two Brocade 3800 (2Gpbs, 16-port) switches. All drives are IBM FC. I'm trying to figure out the best approach to zone these drives. We currently use port zoning on these switches, but as part of this upgrade, we will enable the WWN feature on the L700, and migrate the switches to WWN zoning. Here are my thoughts... Current config: (4) HBAs on the TSM server for tapes On switch1, hba1 has (3) LTO-2 drives hba2 has (4) LTO-2 On switch2, hba3 has (3) LTO-2 plus (3) DLT8K (the DLT's are going away with this upgrade) hba4 has (4) LTO-2 The DLT's are never used, so they have no real impact. I know the hba's with (4) LTO-2 are over-committed, and although I have no direct evidence of a problem, my gut feeling is that it is a bottleneck at times. I have a few FC cards that I can add. There are two open PCI slots on the TSM server... with some more reconfiguration, I can free up two more. This is an HP rp7410 server, and all PCI slots are 66Mhz x 64 bit capable, with throughput of 530MBps so, I can put FC adapters anywhere with no throughput concerns. What I'm considering: - IF I can add four more HBAs (best case - no over-commitment), On each switch, hba1 with (1) LTO-3 and (1) LTO-2 hba2 with (1) LTO-3 and (1) LTO-2 hba3 with (3) LTO-2 hba4 with (2) LTO-2 -If I can only add two HBAs (less reconfiguration), On each switch, hba1 with (1) LTO-3 and (1) LTO-2 hba2 with (1) LTO-3 and (2) LTO-2 (44 MBps over-commitment) hba3 with (4) LTO-2 (24 MBps over-commitment) My over-commitment calculations are based on 2048 Mbps / 8 bits per byte = 256 MBps throughput per switch port. I'm wondering if there are any issues with zoning LTO-2 and LTO-3 to the same HBA... will either one monopolize the port, or will play nice together? Any comments are greatly appreciated! Regards, Robin Sharpe Berlex Labs 973-487-5686
Re: server restore behavior
AFAIK, it's first come first serve. - The different restore sessions will not talk to each other, and the TSM server will not coordinate their resource usage, so it's not A. - Since each session will compete for drives and get the next one available when they get up to bat, it's not B. - And since they all have resourceutilization 4, it's not C. In fact, I'd think it's quite unpredictable, because each client may have differing hardware capabilities, other network traffic will influence it, and what each client is restoring will affect how quickly it gets to the tape mount request(s). It would be nice, though, if the TSM server did coordinate the active sessions. Even nicer would be a facility to define a restore plan, assigning priorities and weights. I suppose you could hack something together with some fancy scripting and/or using an external scheduler like Control-M... but seems like it would be a lot of work. Robin Sharpe Berlex Labs Troy Frank [EMAIL PROTECTED] WISC.EDU To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU server restore behavior 08/18/2006 12:22 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU This is more of a curiosity question than a problem. In a multiple-client restore scenario where you start up say 5 restores at once from different nodes (with 4 tape drives, and resourceutilization set to 4 on nodes), how does the server process the request? Technically, data from all 5 nodes are probably on a lot of the same tapes. Does it A) mount each tape exactly once, getting all data for all running restores off that tape before unmounting. B) Process the restores relatively serially for each node, giving each all 4 drives until completed. Unmounting/remounting the same tapes multiple times. C) Only give each node 1 tape drive to work with, which will effectively ellicit behavior very similiar to option B. Or does it do something different than any of these? Confidentiality Notice follows: The information in this message (and the documents attached to it, if any) is confidential and may be legally privileged. It is intended solely for the addressee. Access to this message by anyone else is unauthorized. If you are not the intended recipient, any disclosure, copying, distribution or any action taken, or omitted to be taken in reliance on it is prohibited and may be unlawful. If you have received this message in error, please delete all electronic copies of this message (and the documents attached to it, if any), destroy any hard copies you may have created and notify me immediately by replying to this email. Thank you.
Re: Detect whether L700 tape library unavailable
The path definition is the only TSM reference I could find that has a library status: ONLINE=YES or NO. Example: ANS8000I Server command: 'q path' Source Name: TSM_WAYNE_LM Source Type: SERVER Destination Name: WATL26 Destination Type: LIBRARY On-Line: Yes I'm not sure if TSM will set it off line if there is a problem. Another approach is to use the lbtest utility that comes with TSM, but I don't know if I'd want to be firing that up automatically it could interfere with TSM activities. BTW, we also have an L700. Ours is attached to an HP rp7410, running HP-UX 11i. We have 14 LTO2 drives, and 4 LTO3's on order. Robin Sharpe Berlex Labs Martin, Roy J [EMAIL PROTECTED] OMTo Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Detect whether L700 tape library unavailable 08/11/2006 11:46 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU We have a TSM server (5.2.2) running on a Solaris box with a SCSI attached STK L700 tape library. TSM interfaces directly with the tape library. Does anyone have any relatively simple method to detect whether the tape library becomes unavailable? I was thinking of enabling some events, have them directed to a log file have a monitoring tool watch for them, but it seems like it would be difficult to identify all of them. Is there some simply way to achieve this? Perhaps query drive or something on a regular basis would generate a particular error if the library was down. I suppose a script could just see if q drive returns valid response within some timeframe. Any thoughts or experience doing this? Roy J. Martin Global Client Engineering GM (BUR group) mailto:[EMAIL PROTECTED]
Re: Detect whether L700 tape library unavailable
Thanks for the tip Len. I don't use show commands in scripts, because A) we're warned that they are unsupported and subject to change without notice, and B) they're a pain to parse. But that does indeed show the library status. As for the ANR8440 msg, that would only happen during TSM startup, woudn't it? When it tries to initialize the library? Off-topic (a little) -- another thing I'd like to be able to see is what tapes (if any) are in the import/export slots. The 'show slots' command lists what slots exist, but not what tapes are in them. Some libraries provide a telnet interface that would probably work, but not the L700 (sigh). Robin Sharpe Berlex Labs Len Boyle [EMAIL PROTECTED] M To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: Detect whether L700 tape library unavailable 08/17/2006 12:38 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Hello Robin, There is an undocumented command show library that has the online status. See the archives for more info on undocumented commands. Here is the partial output from the show library command and the query path command. Autolib is an stk 9710 which was not connected. But at least with a connected library the show library command will show the library as offline if it can not talk to it. You do get the following message in the tsm log for an unconnected library 08/09/2006 09:09:05 ANR8440E Initialization failed for SCSI library AUTOLIB; will retry in 2 minute(s). len - tsm: ADSMNT04show library MMSV-libList: head=03059FB8, tail=0305A5E0 Library AUTOLIB (type SCSI): reference count = 0, online = 0, borrowed drives = 0, update count = 0 basicCfgBuilt = 1, libInfoBuild = 0, definingPathToLibrary = 0 addLibPath = 0, driveListBusy = 0 libvol_lock_id=0, libvolLock_count=0, SeqnBase=0 library extension at 03053960 autochanger list: dev=lb0.0.0.3, busy=0, online=0 Drive detail and second library output deleted. --- tsm: ADSMNT04q path Source Name Source Type Destination Destination On-Line NameType --- --- --- --- --- ADSMNT04SERVER AUTOLIB LIBRARY Yes ADSMNT04SERVER TAPE1 DRIVE Yes ADSMNT04SERVER TAPE2 DRIVE Yes ADSMNT04SERVER BORGLIBRARY Yes ADSMNT04SERVER TAPEB1 DRIVE Yes ADSMNT04SERVER TAPEB2 DRIVE Yes For others: Has anyone asked IBM to move the function of the show library command to a documented command. len -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Robin Sharpe Sent: Thursday, August 17, 2006 11:03 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Detect whether L700 tape library unavailable The path definition is the only TSM reference I could find that has a library status: ONLINE=YES or NO. Example: ANS8000I Server command: 'q path' Source Name: TSM_WAYNE_LM Source Type: SERVER Destination Name: WATL26 Destination Type: LIBRARY On-Line: Yes I'm not sure if TSM will set it off line if there is a problem. Another approach is to use the lbtest utility that comes with TSM, but I don't know if I'd want to be firing that up automatically it could interfere with TSM activities. BTW, we also have an L700. Ours is attached to an HP rp7410, running HP-UX 11i. We have 14 LTO2 drives, and 4 LTO3's on order. Robin Sharpe Berlex Labs Martin, Roy J [EMAIL PROTECTED] OMTo Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Detect whether L700 tape library unavailable 08/11/2006 11:46 AM Please respond to ADSM: Dist Stor
Re: Detect whether L700 tape library unavailable
Gerald, That's a cool script... I've never quite gotten that interface working in batch. The only thing I would worry about is what if TSM tries to open the library while the script has it open? Will it go offline? I guess having it in a script would be less exposure than an interactive run (of lbtest) since it runs at CPU speed. BTW, our L700 (we have two actually, one here in NJ and one in CA) is really an HP 20/700, so it doesn't need the dongle for the web interface to work. Not sure if there are any functional differences, as I've never used the real StorageTek web page. We went to StorageTek when we wanted to upgrade from DLT to LTO and HP could/would not do it. So it's not purple, but beige. Regards, Robin Sharpe Berlex Labs Gerald Michalak [EMAIL PROTECTED] COM To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: Detect whether L700 tape library unavailable 08/17/2006 03:37 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU The L700 does have an optional personality module which allows the library to be connected to the network. You can then open a browser to this address and see the whole library, slots, drives, i/o ports, errors. We have it on all our L700 libraries. Also, I've created a Perl script which uses the lbtest command to get the list of tapes in the i/o slots. ( see below ) It may not be pretty but it works. I use this list in scripts to checkin offsite tapes and load/label new scratch tapes. Gerald Michalak TSM - Certified V5 Administrator == lib_cap_inv.pl == #!/usr/bin/perl # lib_cap_inv.pl # open (LIBV, /usr/tivoli/tsm/devices/bin/lbtest -f lbtest.in -d /dev/lb0 -o /tmp/lbinv.out |) || die( * * * Library Busy ); close (LIBV); `rm -f /adsm_restore/lib_cap_inv`; open (INV,/tmp/lbinv.out); open (CAPINV,/adsm_restore/lib_cap_inv); while (INV) { $line=$_; chop($line); ($type,$dummy,$dummy,$dummy,$slot,$stat ) = split( ,$line); if ( $type eq Import and $stat eq FULL) ) { $line=INV; $line=INV; $line=INV; chop($line); ($dummy,$dummy,$f1,$f2,$f3 ) = split( ,$line); print CAPINV $f1 \n; print slot=.$slot. f1=.$f1. f2=.$f2. f3=.$f3.\n ; } } close (INV); close (CAPINV); exit; == lbtest.in == command open $D command return_elem_count command return_lib_inventory_all command close
Re: TSM Blocks Device Files in HP-UX after using it...
Hi Bernaldo, We also run TSM on HP-UX. I don't recall ever seeing that exact message, but my guess is you either have a hardware problem with those two devices (or something along the path to those devices... are they on the same FC adapter or SCSI bus?), -- OR -- HP-UX has renumbered the device files. I have had that happen, and I'm not sure why... To be sure, you should do an 'ioscan -fnC tape' and compare it to the files in /dev/rmt ('ls -l /dev/rmt'). If the devices have been givien new instance numbers, I have found that the easiest, surest fix is to delete the drives and paths from TSM and redefine them. You should also remove any old unused special files from /dev/rmt by doing 'rmsf -a /dev/rmt/23m' (for example). That command will remove the definition from the kernel and all of the special files that were associated with that device... 23m, 23mn, 23mb, 23mnb. HTH Regards, Robin Sharpe Berlex Labs Bernaldo de Quiros, Iban 1 iban.bernaldodeq To [EMAIL PROTECTED]ADSM-L@VM.MARIST.EDU Sent by: ADSM:cc Dist Stor Manager Subject [EMAIL PROTECTED] TSM Blocks Device Files in HP-UX .EDU after using it... 07/31/2006 05:27 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Hi Everyone !! Did anyone got this error on TSM ¿? any information or GOTCHA's ¿? 07/28/06 12:54:47 ANRD mmsext.c(1360): ThreadId70 Unable to obtain model type for '/dev/rmt/25m', rc = 30 Callchain follows: 00d18e6f outDiagf+0x25f - 00e7005b ExtMountVolume+0xe53 - 0088fbff MmsMountVolume+0x3ff - 00e0a0c7 PvrGtsOpen+0x9a7 - 00e5fa07 TapiOpen+0x247 - 0083ce1f AgentThread+0x7d7 - 0012edb3 StartThread+0x17b - 0006b47b *UNKNOWN* - (SESSION: 506) 07/28/06 15:58:00 ANRD mmsext.c(1360): ThreadId60 Unable to obtain model type for '/dev/rmt/23m', rc = 30 Callchain follows: 00d18e6f outDiagf+0x25f - 00e7005b ExtMountVolume+0xe53 - 0088fbff MmsMountVolume+0x3ff - 00e0a0c7 PvrGtsOpen+0x9a7 - 00e5fa07 TapiOpen+0x247 - 0083ce1f AgentThread+0x7d7 - 0012edb3 StartThread+0x17b - 0006b47b *UNKNOWN* - (SESSION: 31) After appearing this issues on actlog, TSM do not release the device file, so device file remains busy. The only solution that I find at the moment is to halt the TSM server to release device files. When I re-start the TSM it could access the device files normally until this issue occurs another time... OS VERSION HP-UX 11.00 TSM 5.2.6.3 Using HP-UX Generic Driver to access tapes (stape). Working with GRESHAM EDT 7.4.0.5. Any help will be appreciated. Thanks in Advance, Regards, Ibán Bernaldo de Quirós Márquez Technical Specialist cell: + 34 659 01 91 12 Sun Microsystems Iberia
Re: TapeAlerts on LTO2 drives, tapes
Thanks Jurjen and Richard, Sorry for omitting the subject.. I forgot to mention in my original post that STK guy upgraded all drives to the latest firmware a couple of weeks ago... we are now at 53Y2. I just discovered that one drive is at 5AT0... and it is the one with the highest number of occurrences of that tapealert. We also upgraded the library code to 3.11.02. I wonder if these levels could be buggy... We discovered the CM problem a year ago. But, we only noticed it when attempting to read a corrupted tape, and we got a different tapealert. I always assumed that there was no indication when the corruption occurred, only when it was found. Is this tapealert new to TSM 5.3 (vs. 5.2 which we were at last year)? It's certainly better to detect the problem when it happens than maybe years later! We handled the CM problem by setting up an Openview monitor that started an audit of each volume that got the tapealert. When we started, we had maybe 20 audits a day. Over several months, the audits decreased, and now we rarely have one. We probably still have some corrupted tapes in the inventory, as many have not been mounted in years. I'm not sure if we can use tapeutil in the HP-UX environment, but I didn't like that approach anyway because it's a manual process. Robin Sharpe Berlex Labs Jurjen Oskam [EMAIL PROTECTED] S.ORG To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: TapeAlerts on LTO2 drives, tapes 06/29/2006 03:48 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU On Wed, Jun 28, 2006 at 02:23:33PM -0400, Richard Sims wrote: You might experiment with one of your drives, boosting it to the latest microcode level (if not already there) and see if amelioration occurs. Back in 2005 we also had problems with our IBM LTO2 drives in a 3584 library. In the end we found out that these were all microcode-related. After upgrading to a fixed microcode level, we still kept seeing problems with some tapes. What happened was that the chip in the cartridge contained 'bad' information (due to the drive microcode problems earlier on) and this causes weird behaviour (TapeAlerts) even after we installed correct microcode. The fix for this was to use 'Space to End of Data' in tapeutil: this apparently clears or reinitializes the chip in the tapes. (Even now, we sometimes get a TapeAlert This tape is not data-grade on a rarely used volume.) -- Jurjen Oskam Savage's Law of Expediency: You want it bad, you'll get it bad.
[no subject]
Hello TSMers, I've been getting messages like the following on several drives and for several cartridges: 06/28/06 10:16:58 ANR8948S Device /dev/rmt/16m, volume 205975 has issued the following Critical TapeAlert: The tape just unloaded could not write its system area successfully: 1. Copy data to another tape cartridge. 2. Discard the old cartridge. (SESSION: 4514) 06/28/06 10:16:58 ANR8948S Device /dev/rmt/16m, volume 205975 has issued the following Critical TapeAlert: The operation has failed because the media cannot be loaded and threaded. 1. Remove the cartridge, inspect it as specified in the product manual, and retry the operation. 2. If the problem persists, call the tape drive supplier help line. (SESSION: 4514) We have been having more and more drive and tape problems over the last six months or so, after a couple years of solid performance. I'm trying to figure out the root cause or causes. I suspected drives, because I also saw write errors on specific drives... we swapped them out, and it seemed better for a while... now I don't see the write errors, but I see the messages above. Some clues... these seem to be occurring only on recently purchased, new tapes. We did change vendors about a year ago to get better pricing. I have to collect some stats on how many tapes have problems, but my guess is only a dozen or so among the 4000 tapes we purchased over the past year from this vendor. The only other thing I can think of is changing the library to autolabel. Oh, and we upgraded from TSM 5.2 to 5.3 in April '06. Is this a tape problem (bad batch or low quality), or a drive or library problem? My environment: TSM 5.3.2.0 on HP-UX 11i. Big server (HP rp7410, 8 CPU, 12GB RAM, HP XP512 RAID5 disk) Five TSM servers on this box -- one library manager, four data handlers. The errors above are in the library manager's activity log. STK L700 library w/ six DLT7000 (not being used) and 14 IBM LTO2. 618 slots, library is usually full. We have about 3000 tapes onsite. Library is doing tape cleaning, one universal cleaning tape in the library. Sometimes I see Tapealerts saying the cleaning tape is not data grade... was TSM trying to write to it? Should we let TSM do the cleaning instead? TIA Robin Sharpe Berlex Labs
Re: I need some clarification
If you need to GUARANTEE data retention for 45 days, set the parameters as follows: Versions Data Exists = NOLIMIT Versions Data Deleted = NOLIMIT Retain Extra Versions = 45 Retain Only Version = 45 Why? The Version Data Exists and Deleted parameters specify a number of versions. If you, or someone else, runs more than one selective or absolute mode backup per day, you may not get as many days retained as you intended. The Retain parameters specify the number of days to keep inactive versions. Robin Sharpe Berlex Labs Lepre, James [EMAIL PROTECTED] COM To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: I need some clarification 06/07/2006 06:17 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Hello Everyone, I need some clarifications. What do the following parameters mean, in basic terms..I need to be able to recover for instance from today 45 days ago, so how would I set up these parameters I have ask tsm support and I cannot get a clear answer, so hopefully someone can help me. Versions Data Exists Versions Data Deleted Retain Extra Versions Retain Only Version Thank you in advance James --- Confidentiality Notice: The information in this e-mail and any attachments thereto is intended for the named recipient(s) only. This e-mail, including any attachments, may contain information that is privileged and confidential and subject to legal restrictions and penalties regarding its unauthorized disclosure or other use. If you are not the intended recipient, you are hereby notified that any disclosure, copying, distribution, or the taking of any action or inaction in reliance on the contents of this e-mail and any of its attachments is STRICTLY PROHIBITED. If you have received this e-mail in error, please immediately notify the sender via return e-mail; delete this e-mail and all attachments from your e-mail system and your computer system and network; and destroy any paper copies you may have in your possession. Thank you for your cooperation.
Re: the MACRO command, but on the server?
I don't think it matters where you run dsmadmc from, once you issue the macro command, TSM will look in the current directory, or the absolute path if given, on the TSM server (not on the client machine). Also, if you redirect output from a query or select command, it will be stored on the TSM server. Is that what you are trying to do? Robin Sharpe Berlex Labs Jim Zajkowski [EMAIL PROTECTED] U To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU the MACRO command, but on the server? 05/30/2006 10:26 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Hey folks, Is there a way to have TSM execute a command file, but one stored on the server? Here's what I'm trying to do: I want to be able to do a MOVE DRM from my desktop (where I have dsmadmc running locally), have it generate the checkin script, and then execute that script. The reason is that we're using named private volumes for our offsites, but scratch tapes for our db backups, and I want to issue the correct checkin status. Why not just execute dsmadmc on the server? I could, and I probably will, but I'm eventually planning on constructing a click-and-drool applet to make this easier for my other admins when I'm on vacation, and I prefer location independence -- the app may run on a different box entirely. Any ideas? --Jim
Re: multiple instance recommendations
Steve, I envy your memory! Unfortunately, we are still under a permanent hold notice due to pending litigations. And, yes, I'd say that is the primary reason why our database grew so large. As large as it is, the size itself was never a big problem, except that DB restores in Disaster Recovery Tests took about 5+ hours. In our last test, just after the split, it took only three minutes to restore our new 12GB DR TSM DB. Many folks with large TSM DBs seem to have problems with expirations running too long but we haven't done one for over two years now, so wasn't an issue for us. I suppose we may have had some performance issues with the large DB, but it's difficult to quantify. Anyway, the biggest problem we have with the permanent retention is with tape consumption. We now have about 6,000 LTO2 volumes... about half onsite. We have a STK L700 library with 618 slots, so we are constantly moving media out to make room for new tapes and returns from the vault. That's compounded because last October, our IT group moved to a new office about eight miles from our data center. We now have some of the PC techs in the data center doing our daily moves, but sometimes we have to travel to handle problems, emergency mount requests, etc. Now that we've split to five TSMs (the old one, the new LM, one for Lotus Notes, one for Windows, and one for DR and the remaining Unix and Oracle) the daily tape moves have become more complicated... but we're getting a handle on it. The key was to create a server group and use it to broadcast all the DRM commands to all servers. Regards, Robin Sharpe Berlex Labs Steven Harris [EMAIL PROTECTED] IS.INFO To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: multiple instance recommendations 05/28/2006 08:56 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Robin, If I recall correctly a while back you were stuck with having to keep everything forever. Is that still the case and is that why your database got to 530GB? If it is still the case I'd be interested in how you handle this, problems encountered and so on. Regards Steven Harris AIX and TSM Administrator, Brisbane Australia On 27/05/2006, at 12:43 PM, Robin Sharpe wrote: I just did this about a month ago... first the why: My single TSM database had grown to its architectural limit: 530GB. I had no choice other than splitting TSM into multiple instances (well, I could have frozen the big TSM and started over... and we actually ended up doing that also... but multiple instances seemed like a good approach for a number of reasons). We didn't have the budget for new servers, and our TSM server is pretty hefty - an HP rp7410 w/ 8 CPUs and 12GB RAM. All DB and log volumes are on an HP XP512 (Hitachi Lightning). Now the how: Use the same TSM install directory as per the install guide. Use the same host name/IP address for each instance, but assign a unique set of ports (TCPPort, HTTPPort, SHMPort) for each instance. Then assign the appropriate port in the dsm.sys/dsm.opt in each client. And of course register the clients in the corresponding TSM instance. As far as tape access, using one server means you only need to have the drives physically connected to that server (unless you're doing LAN free backups)... and this may enable sharing your SCSI connected drives, although I haven't tried that... we have two SCSI connected DLT drives, but we almost never use them anymore. One other recommendation, which I did... do not run all of your TSM instances as root (you are on Unix , right?). Doing so, if you do a ps -ef|grep dsmserv command, you can't tell which process belongs to which server without doing more research .. looking up the PIDs. Instead, create a unique user for each TSM and run it under that user. Make sure you chown all of your config files and DB/log/stgpool volumes to that user. This will also prevent the wrong TSM from accessing another TSM's storage... this can be easy to do if you use raw logical volumes as we do. And now, your ps output will show the user for each dsmserv. It's actually quite easy to set up an additional instance once you get all the steps down... I'll share my approach in more detail if anyone is interested... Regards, Robin Sharpe Berlex Labs ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 05/22/2006 05:28:32 PM: Other than the obvious hardware cost savings, I don't really see the advantage of multiple instances on the same
Discovered a new secret command -- anyone know what fetch does?
So, the other day, we replaced two LTO2 drives in our STK L700... new WWNs, so had to delete redefine the drives paths. I have four servers sharing the library, so that's four sets of path definitions... to make it easier I wrote up some macros to delete the paths for each drive then define them. Well, due to an editing error, the de in the first def path command got deleted... so the command issued was f path TSM rejected this in an interesting way (recreated here): IBM Tivoli Storage Manager Command Line Administrative Interface - Version 5, Release 2, Level 0.0 (c) Copyright by IBM Corporation and other(s) 1990, 2003. All Rights Reserved. Session established with server TSM_WAYNE_LM: HP-UX Server Version 5, Release 3, Level 2.0 Server date/time: 05/26/06 22:13:17 Last access: 05/26/06 22:13:09 ANS8000I Server command: 'f path' ANR2000E Unknown command - FETCH PATH. ANS8001I Return code 2. ANS8002I Highest return code was 2. TSM expanded the f to FETCH... I just thought it was amusing... Robin Sharpe Berlex Labs
Re: multiple instance recommendations
I just did this about a month ago... first the why: My single TSM database had grown to its architectural limit: 530GB. I had no choice other than splitting TSM into multiple instances (well, I could have frozen the big TSM and started over... and we actually ended up doing that also... but multiple instances seemed like a good approach for a number of reasons). We didn't have the budget for new servers, and our TSM server is pretty hefty - an HP rp7410 w/ 8 CPUs and 12GB RAM. All DB and log volumes are on an HP XP512 (Hitachi Lightning). Now the how: Use the same TSM install directory as per the install guide. Use the same host name/IP address for each instance, but assign a unique set of ports (TCPPort, HTTPPort, SHMPort) for each instance. Then assign the appropriate port in the dsm.sys/dsm.opt in each client. And of course register the clients in the corresponding TSM instance. As far as tape access, using one server means you only need to have the drives physically connected to that server (unless you're doing LAN free backups)... and this may enable sharing your SCSI connected drives, although I haven't tried that... we have two SCSI connected DLT drives, but we almost never use them anymore. One other recommendation, which I did... do not run all of your TSM instances as root (you are on Unix , right?). Doing so, if you do a ps -ef|grep dsmserv command, you can't tell which process belongs to which server without doing more research .. looking up the PIDs. Instead, create a unique user for each TSM and run it under that user. Make sure you chown all of your config files and DB/log/stgpool volumes to that user. This will also prevent the wrong TSM from accessing another TSM's storage... this can be easy to do if you use raw logical volumes as we do. And now, your ps output will show the user for each dsmserv. It's actually quite easy to set up an additional instance once you get all the steps down... I'll share my approach in more detail if anyone is interested... Regards, Robin Sharpe Berlex Labs ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 05/22/2006 05:28:32 PM: Other than the obvious hardware cost savings, I don't really see the advantage of multiple instances on the same hardware. (I haven't decided yet if we would use one beefy server or two medium servers.) If you load up multiple instances on the same server, do you give them different IP interfaces to make distinguishing between them in client configs and administration tools easier? Tape access-wise, is there a hardware advantage putting multiple instances on the same system? Any recommendations on any of this? Your help is appreciated. Dave
[no subject]
Dear colleagues, It's time for us to split our TSM into several new instances because our database is now just too large -- 509GB -- and still growing. My initial plan is to create five TSMs - four plus a library manager - on the existing server (an 8-way, 12GB HP rp7410 with 15 PCI slots). This is cost effective since no additional hardware or license is needed - just lots of SAN disk for the databases, which we have available. But, I've been thinking what do you think about the following: A more creative approach is to place the new TSM servers on existing large clients. This has several advantages: - eliminates need to acquire new servers, saving physical room, power and cooling requirements, additional maintenance. - client benefits by sending its backup to local disk using shared memory protocol. Eliminates potential network bottleneck. - Client sends data to tapes using library sharing; no need for storage agent. - Use of local disk eliminates the need for SANergy - heavy clients pay for their usage by providing backup services for smaller clients. There are also some concerns (not necessarily disadvantages): - May require CPU, memory, and/or I/O upgrades (still cheaper than buying a server) - TSM operation may impact client's primary app. Can be controlled by PRM on HP-UX. - Incurs licensing cost. Thanks for any insights Robin Sharpe Berlex Labs
TSM Server Hosting - dedicated vs. shared
Orville, Thanks for your thoughts. We do use Control-M for all of our scheduling in the Unix environment, and are moving towards Windows deployment too. I am surprised, though, about your comment on licensing. I thought each TSM server instance on a separate physical server needed a license (per processor). Is this not true? Is it a new policy? Robin Sharpe Berlex Labs |-+--- | | Orville Lantto | | | [EMAIL PROTECTED]| | | SHOUSE.COM | | | Sent by: ADSM: Dist| | | Stor Manager | | | [EMAIL PROTECTED]| | | U | | | | | | | | | 03/13/2006 01:40 PM | | | Please respond to | | | ADSM: Dist Stor| | | Manager| | | | |-+--- | | | | | |To: ADSM-L@VM.MARIST.EDU | |cc: | |Subject: | |Re: | | The approach is valid and can reap significant backup/restore time benefits for the clients. Two points: 1) No new licensing cost are involved. TSM is licensed by the environment, not the number of TSM servers. 2) Consider the complexity of resources scheduling between many servers. Most sites have a limited number of tape drives and contention can be a bear to schedule out with so many independent servers and their separate schedulers. An external admin scheduling utility may be needed. Orville L. Lantto Glasshouse Technologies, Inc. From: ADSM: Dist Stor Manager on behalf of Robin Sharpe Sent: Mon 3/13/2006 11:03 AM To: ADSM-L@VM.MARIST.EDU Subject: [ADSM-L] Dear colleagues, It's time for us to split our TSM into several new instances because our database is now just too large -- 509GB -- and still growing. My initial plan is to create five TSMs - four plus a library manager - on the existing server (an 8-way, 12GB HP rp7410 with 15 PCI slots). This is cost effective since no additional hardware or license is needed - just lots of SAN disk for the databases, which we have available. But, I've been thinking what do you think about the following: A more creative approach is to place the new TSM servers on existing large clients. This has several advantages: - eliminates need to acquire new servers, saving physical room, power and cooling requirements, additional maintenance. - client benefits by sending its backup to local disk using shared memory protocol. Eliminates potential network bottleneck. - Client sends data to tapes using library sharing; no need for storage agent. - Use of local disk eliminates the need for SANergy - heavy clients pay for their usage by providing backup services for smaller clients. There are also some concerns (not necessarily disadvantages): - May require CPU, memory, and/or I/O upgrades (still cheaper than buying a server) - TSM operation may impact client's primary app. Can be controlled by PRM on HP-UX. - Incurs licensing cost. Thanks for any insights Robin Sharpe Berlex Labs
Re: Alleged LTO fragility
I don't know about LTO1, we have LTO2. They are fragile, but we have not had problems as others have described... I have dropped a few myself, and I can testify that a drop from as little as a foot can ruin an LTO2 cart. I have also dropped them from waist-height (about 3 feet) with no damage. The problem I have seen is that if the cart lands on the corner where the tape comes out, there is a thin point in the plastic where the leader pin slips in... this is a weak point that can bend which causes the cart to widen slightly at that point... but it's enough to prevent the cart from loading. If the damage is bad enough, the pin will fall out of its place. I have 7 or 8 carts (out of 5600) that have failed this way. Interestingly, I don't recall having any come back from the vault damaged... and we do not use the jewel cases, we just load the carts into red turtle cases for transport. Pretty nice cases, though... foam lined, they hold the carts pretty firmly. We can fit about 36 carts in one case. We usually ship 25-30 offsite each day. Robin Sharpe Berlex Labs Tab Trepagnier [EMAIL PROTECTED] AITRAM.COM To: ADSM-L@VM.MARIST.EDU Sent by: ADSM: cc: Dist Stor Subject: Manager Alleged LTO fragility [EMAIL PROTECTED] .EDU 02/10/2006 12:51 PM Please respond to ADSM: Dist Stor Manager This is a tape media question. We are considering moving from DLT8000 to LTO-1 for offsite tape storage. Our local IBM guys told us that LTO tapes should not be dropped more than 1/2 inch or they risk being damaged. If you've ever watched a vault vendor handling tapes, you know that 1/2 is pretty optimistic. So, is that true? Would our offsite data be at risk if our vault vendor dropped an LTO cartridge a whole inch? Thanks. Tab Trepagnier TSM Administrator Laitram, L.L.C.
Re: Antwort: Re: Trouble when TSM-database reaches 100Gb?
Not only do I not run expirations, I have also made copies of each policy set and changed all retention parameters to NOLIMIT. I made copies so that I can easily return to normal when this situation is resolved. |-+--- | | TSM | | | [EMAIL PROTECTED]| | | | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 09/06/2005 02:17| | | AM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- --| | | | | |To: ADSM-L@VM.MARIST.EDU | |cc: | |Subject: | |Antwort: Re: Trouble when TSM-database reaches 100Gb? | --| hello Robin, you wrote, that you can not delete any file yet, so you suppress expiration. but, when a file reaches the threshholds for versions exists or the other three parameters in copygroups for backups (archives similar), it is lost for restores immediately. Expiration is not the key in this case! with best regards stefan savoric
Re: Trouble when TSM-database reaches 100Gb?
Hans, I don't believe that 100GB is a magic limit. Just doesn't make sense. The TSM server is a very complex thing; there are many factors affecting performance, as others have stated. For the record, we run a 370GB TSM on an HP rp7410 with (8) 875MHz processors and 12GB RAM. This is a Unix (HP-UX) machine. On the weekends, we typically have over 100 backup sessions running at some times, and during those periods, the CPU utilization (all 8 processors) goes to nearly 100%. But, the TSM DB just runs and runs. Others have highlighted expiration problems; due to circumstances beyond my control, we cannot delete anything at the moment, so I don't run any expirations at all. When we start them again (someday soon, I hope), I suspect they will run a long time ):. Although the TSM DB size itself does not appear to be a problem, it does make DB restores (done regularly in Disaster Recovery tests) take quite some time. It can take up to a full day to get my TSM ready to start client restores in DR tests. For that reason, among others, I am planning to split to several TSM DB's... but they will all be hosted on the same physical server. regards, Robin Sharpe Berlex Labs Hans Christian Riksheim [EMAIL PROTECTED] To: ADSM-L@VM.MARIST.EDU Sent by: ADSM: cc: Dist StorSubject: Manager Trouble when TSM-database reaches 100Gb? [EMAIL PROTECTED] T.EDU 09/01/2005 04:16 AM Please respond to ADSM: Dist Stor Manager Hello, we are experiencing throughput problems on our TSM-installation. Apart from he obvious that we have too few tape drives, bottlenecks in our LAN and an old AIX-box, we got a suggestion to add another TSM-server. The reason was that the performance degrades when the TSM-database reaches about 100Gb in size. We find it a little troublesome to add a new TSM server every time we reach 100 Gb, we would rather just buy us a new AIX-box which is 5 times faster. So I ask you if this 100 Gb limitiation really applies. What is your experience with this? Server version 5.2.3.3 Ca. 100 clients. Best regards Hans Chr. Riksheim
Re: More tapes used than necessary
Need a little clarification here Are your clients backups going to a disk pool first or directly to SQL_TAPEPOOL? Is SQL_TAPEPOOL your primary tape pool, or a copy pool? The normal sequence of events if you're backing up to disk is something like this: - Clients backup to the disk pool throughout the day. - Sometime later, you do a backup stgpool to backup the disk pool to the tape COPY pool. - Then you do a migration of the disk pool to the tape PRIMARY pool (SQL_TAPEPOOL in your case?) (this could also happen throughout the day based on threshold value) - Then you do backup stgpool of the tape PRIMARY pool to the tape COPY pool. - Then you send the copy pool volumes offsite. If your problem is occurring during the migration, check the DISK pool definition and see if migration processes is set greater than 1. If so, TSM will need that many tapes to write to. Robin Sharpe Berlex Labs |-+--- | | ML_SPAM | | | [EMAIL PROTECTED]| | | A.COM | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 08/03/2005 07:09| | | AM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- --| | | | | |To: ADSM-L@VM.MARIST.EDU | |cc: | |Subject: | |More tapes used than necessary | --| Hi *SMers, TSM 5.2.1 server W2K sp4 3583 L72 I can't get my head around why this is happening... A number of SQL clients back up to the SQL_TAPEPOOL, with collocation turned off. However, we are only storing 68 MB from all the sql clients so I would expect all that data to fit on one tape, which it does. But for some reason TSM wants to pick up two scratch volumes everyday to backup the SQL diskpool to - not the already defined volume. Eachmorning I find myself performing a 'move data' on the two newly defined vols just to move the data back to the slowly accumilating volume. One thing that occured to me was that the volume had about 37% reclaimable space - and if the files which had expired were at the end of the volume, would TSM be able to locate the correct position to start writing from? That's the only reason I can think why this is happening. Can any one else shine a bit of light on this problem? Many Thanks, Matthew Aviva plc Registered Office: St. Helen's, 1 Undershaft, London EC3P 3DQ Registered in England Number 02468686 www.aviva.com This message and any attachments are confidential. If you are not the intended recipient, please telephone or e-mail the sender and delete this message and any attachment from your system. Also, if you are not the intended recipient you must not copy this message or attachment or disclose the contents to any other person.
Re: Announcing IBM Support Assistant for TSM
Darrius, I installed the assistant, and downloaded the TSM plugin, extracted the files to the plugins subdirectory according to the instructions, but TSM is not showing up anywhere within the assistant. I only see the Websphere plugin. Can you clarify the plugin install procedure? Thanks Robin Sharpe Berlex Labs |-+--- | | Darrius Plantz | | | [EMAIL PROTECTED]| | | AHOO.COM | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 07/12/2005 10:27| | | AM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- ---| | | | | |To: ADSM-L@VM.MARIST.EDU | |cc: | |Subject: | |Announcing IBM Support Assistant for TSM | ---| The TSM development staff has just finished the initial implementation of a tool that will assist Tivoli Storage Manager admins with their support issues. It is called IBM Support Assistant ( ISA) and is available via the download section at the TSM Support page: http://www-306.ibm.com/software/sysmgmt/products/support/IBMTivoliStorageManager.html ISA is a cross-product extensible client application that increases your capacity for self-help by making it easier to access support resources. ISA allows you to add plug-ins for the IBM products you use, creating a customized support experience for your configuration. The Tivoli Storage Manager Server product plug-in enhances the Search, Education, Support Links and Service component of IBM Support Assistant. As more plugins become available for TSM we will announce them here. __ Do You Yahoo!? Tired of spam? Yahoo! Mail has the best spam protection around http://mail.yahoo.com
Quantum PX720/ HP ESL 712e
Is anyone using either of these libraries? The HP ESL 712e is a re-badged Quantum PX720 which we are considering purchasing. Anyone using expansion frames? We are looking at a two-frame configuration which HP has just finished testing with TSM. TIA Robin Sharpe Berlex Labs
Re: Very slow restores (days), hours to locate files
Thanks for the tip, Bill. But in our environment, (618 tapes in the L700 library, plus over 1000 on a rack), that would take quite a long time... BTW, we think it's a firmware problem... We are running 38D0 on our LTO-2 drives. I just spoke to my STK CE, and he said the latest STK has certified is 4C60, which we will install tomorrow. We're also upgrading the library code from 3.07.00 to 3.09.00. -Robin William Boyer [EMAIL PROTECTED] T.NET To: ADSM-L@VM.MARIST.EDU Sent by: ADSM: cc: Dist StorSubject: Manager Re: Very slow restores (days), hours to locate files [EMAIL PROTECTED] T.EDU 07/07/2005 11:11 AM Please respond to ADSM: Dist Stor Manager TO identify the volumes with the corrupted CM, just do a checkin with CHECKL=YES. This will then issue a tapealert message for a tape that has a bad/corrupted CM. If it's a scratch tape, then the CM will be written as you write to the tape. For tapes with data on them, you can MOVE DATA to another drive and let it go scratch, or use the tapeutil utility to mount the tape and forward to the end of the tape. That causes the drive to re-write the CM index. That is after you've identified the problem that is corrupting your CM chips. It could be firmware, but at one client site of mine, it was a faulty drive. Bill Boyer Some days you're the bug, some days you're the windshield - ?? -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Rejean Larivee Sent: Thursday, July 07, 2005 10:51 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: Very slow restores (days), hours to locate files Hello Robin, the TSM server does not maintain the LTO memory cartridge and would therefore not be the source of the corruption. A corrupted memory cartridge comes from defective media, faulty/dirty hardware/drive or firmware/microcode problem. As others have already recommended, you should consider upgrading the firmware of the LTO drives to take care of past problems with LTO CM. It appears the latest firmware for the LTO GenII drives at this time is 53Y2, for fiber attached drives. You can verify the firmware of your drives using lscfg -vl rmt*. For a list of what is fixed in 53Y2, see here : http://www-1.ibm.com/support/docview.wss?rs=0uid=ssg1S1002360 Have a great day ! Rejean Larivee IBM Tivoli Storage Manager support Robin Sharpe [EMAIL PROTECTED] LEX.COM To Sent by: ADSM: ADSM-L@VM.MARIST.EDU Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: Very slow restores (days), hours to locate files 07/06/2005 04:12 PM Please respond to ADSM: Dist Stor Manager Sorry about the omission, Rich. These restores were started via the Windows GUI. I believe they just selected the C: drive and specified Restore if newer (an option which I don't think is available via the command line!). I believe this created a No-Query Restore, because it did create a Restartable Restore AFAIK there is a one-to-one correspondence (right?) In the meantime, I checked the Technote... Then, I checked my Activity Log for the last 24 hours... and I found 33 LTO volumes that presented the cartridge memory message! So, now I have the smoking gun, and I suppose I could do move data against those volumes, but I suspect there are many more, and I would like to know what's causing the corruption and how to prevent it! If I don't hear anything from the group, I'll open a call with Tivoli. Thanks very much for the information! -Robin Richard Sims [EMAIL PROTECTED] Sent by: ADSM: To: ADSM-L@VM.MARIST.EDU Dist Storcc: Manager Subject: [EMAIL PROTECTED] Re: Very slow restores (days), hours to locate files T.EDU 07/06/2005 10:30 AM Please respond to ADSM: Dist Stor Manager Please, everyone, when posting questions about restorals, give details about the manner in which the restoral was invoked so that we can get a sense of what kind is involved (NQR, Classic) and what is involved. Now... Robin, have a look at IBM Technote 1209563, which I ran across in doing
Re: Very slow restores (days), hours to locate files
The pool in question here is not collocated, so I expected to have lots of tape mounts but the long period of inactivity is what puzzles us. I've also seen notes indicating that the No Query Restore may be the cause... I'll suggest to our Windows guys to try a classic restore. -Robin Connor, Jeffrey P. [EMAIL PROTECTED] To: ADSM-L@VM.MARIST.EDU .NGRID.COMcc: Sent by: ADSM:Subject: Dist Stor Manager Re: Very slow restores (days), hours to locate files [EMAIL PROTECTED] EDU 07/07/2005 04:39 PM Please respond to ADSM: Dist Stor Manager Robin, I hope the LTO firmware resolves your problem. However, I have seen a similar situation for Windows clients in our shop and it was not a tape drive issue. The situation here was that we had a tape stgpool, 3590Ks /3590E1A drives, collocated by node, that reached its maxscratch value. This led to what some folks call imperfect collocation where even though the stgpool is setup to collocate by node, data for more than one node can end up on the same tape. The problem we had with the node intermix in a collocated by node pool showed itself with a situation that sounds similar to yours. We attempted to run a restore of an 8GB Win2k C: drive about 50% full and saw very long delays where nothing appeared to be happening. A tape change would occur, some data would transfer, and then a VERY long pause before a mount request for the next tape. Query Session while tape was mounted showed what your Q SE showed below, session in Run state, zero seconds wait time, but send and recv byte counts remain unchanged. While not as many but similar to you, our incremental backups of the servers C: drive had files spread around a number of tapes. We never determined the root cause of the think time between tape mount requests. We resolved the issue by moving tapes in our stgpool to a new pool with a high MAXSCR value effectively re-collocating the data. All restores ran very happy after that. Sorry I could not provide a root cause of our situation but that's how we addressed it. Just curious but do the 310 tapes you identified also contain data for other nodes and are you using collocation? Jeff Connor National Grid USA -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Robin Sharpe Sent: Wednesday, July 06, 2005 10:10 AM To: ADSM-L@VM.MARIST.EDU Subject: Very slow restores (days), hours to locate files Hi guys, We're having problems restoring some windows servers (W2K)... The servers in question had some disk problems and are being rebuilt, so the Windows admins are restoring the C: drive. It is an 8GB drive and less than 50% used, so only 4GB to restore. It has taken several days to restore. I know one of our problems is that the data is spread over hundreds of volumes (literally... I counted 310 from a volumeusage query). Another problem is that we have an overflowed library, but we have loaded all of the tapes from the Windows storage pool. What I don't understand is why it takes so long to locate a file once the tape is mounted. We have seen the same tape mounted for hours before any data is transferred. Here is an excerpt from a q se f=d of a restore that is running right now: Sess Number: 1,143 Comm. Method: TCP/IP Sess State: Run Wait Time: 0 S Bytes Sent: 670.9 M Bytes Recvd: 58.2 K Sess Type: Node Platform: WinNT Client Name: WANO01 Media Access Status: Current input volume(s): 200658,(2279 Seconds) User Name: Date/Time First Data Sent: Proxy By Storage Agent: This restore has been running for almost 12 hours now (they have been restarting them periodically). There has been NO DATA transferred from that tape in the 38 minutes it has been mounted... I know this from doing an lsof command and looking at the offset which indicates the number of bytes transferred. I know that when I restore a single file, it can be found within seconds of mounting a tape (these are all LTO-2)... so, why does it take so long in this case? Is TSM actually reading the entire tape? If so, wouldn't I see lots of data being transferred? Or is there some kind of SCSI command that allows the drive to read and compare the data it gets? I thought TSM stored actual locations of the files in the DB, so it could quickly find any file (or aggregate) without reading the whole tape... I've been searching the literature, and I can't find any details on this. The TSM server is on HP-UX 11i, IBM LTO-2 drives, fiber attached, in a STK L700 library. Also, my
Very slow restores (days), hours to locate files
Hi guys, We're having problems restoring some windows servers (W2K)... The servers in question had some disk problems and are being rebuilt, so the Windows admins are restoring the C: drive. It is an 8GB drive and less than 50% used, so only 4GB to restore. It has taken several days to restore. I know one of our problems is that the data is spread over hundreds of volumes (literally... I counted 310 from a volumeusage query). Another problem is that we have an overflowed library, but we have loaded all of the tapes from the Windows storage pool. What I don't understand is why it takes so long to locate a file once the tape is mounted. We have seen the same tape mounted for hours before any data is transferred. Here is an excerpt from a q se f=d of a restore that is running right now: Sess Number: 1,143 Comm. Method: TCP/IP Sess State: Run Wait Time: 0 S Bytes Sent: 670.9 M Bytes Recvd: 58.2 K Sess Type: Node Platform: WinNT Client Name: WANO01 Media Access Status: Current input volume(s): 200658,(2279 Seconds) User Name: Date/Time First Data Sent: Proxy By Storage Agent: This restore has been running for almost 12 hours now (they have been restarting them periodically). There has been NO DATA transferred from that tape in the 38 minutes it has been mounted... I know this from doing an lsof command and looking at the offset which indicates the number of bytes transferred. I know that when I restore a single file, it can be found within seconds of mounting a tape (these are all LTO-2)... so, why does it take so long in this case? Is TSM actually reading the entire tape? If so, wouldn't I see lots of data being transferred? Or is there some kind of SCSI command that allows the drive to read and compare the data it gets? I thought TSM stored actual locations of the files in the DB, so it could quickly find any file (or aggregate) without reading the whole tape... I've been searching the literature, and I can't find any details on this. The TSM server is on HP-UX 11i, IBM LTO-2 drives, fiber attached, in a STK L700 library. Also, my DB is huge (314GB), and we are currently (for the last year) unable to delete anything, so we have many versions of volatile files. We are planning to split our environment into several TSMs, and in the short term, our windows admins will start doing weekly selective backups of the C: drives to consolidate active versions on few tapes. Thanks for any thoughts on this Robin Sharpe Berlex Labs
Re: Very slow restores (days), hours to locate files
Sorry about the omission, Rich. These restores were started via the Windows GUI. I believe they just selected the C: drive and specified Restore if newer (an option which I don't think is available via the command line!). I believe this created a No-Query Restore, because it did create a Restartable Restore AFAIK there is a one-to-one correspondence (right?) In the meantime, I checked the Technote... Then, I checked my Activity Log for the last 24 hours... and I found 33 LTO volumes that presented the cartridge memory message! So, now I have the smoking gun, and I suppose I could do move data against those volumes, but I suspect there are many more, and I would like to know what's causing the corruption and how to prevent it! If I don't hear anything from the group, I'll open a call with Tivoli. Thanks very much for the information! -Robin Richard Sims [EMAIL PROTECTED] Sent by: ADSM: To: ADSM-L@VM.MARIST.EDU Dist Storcc: Manager Subject: [EMAIL PROTECTED] Re: Very slow restores (days), hours to locate files T.EDU 07/06/2005 10:30 AM Please respond to ADSM: Dist Stor Manager Please, everyone, when posting questions about restorals, give details about the manner in which the restoral was invoked so that we can get a sense of what kind is involved (NQR, Classic) and what is involved. Now... Robin, have a look at IBM Technote 1209563, which I ran across in doing research yesterday. I recall such long-duration- restores in the past, and as I recall they have involved the factors noted in the Technote. LTO is also known for backhitch delays, so that's another contributor in positioning on tape. Richard Sims
strange problem - missing file on server
Hi All, Has anyone run into this situation before? I was asked to restore a single file on a Unix (HP-UX 11i) server. So I fired up the web client, looked in the specified directory, no file, not active or inactive. The file does exist on the client, with a date stamp of Jan 28 2003, and permissions rwxr-x---. There are other files in the directory with same permissions and old (even older) date stamps, that do exist on the server. We keep logs of our daily backups, so I cthe most recent one, which was run earlier the same day (yesterday)... the directory was backed up, but the file was not. The file's name is bkupscripts... nothing weird, no special characters or backspaces in the name a 'find . -name bkupscripts' command does find the file, as does a 'ls -l' command. Client level is 5.2.0.0, Server is 5.2.4.2 (although client reports it as 5.2.3.2... don't know why) TSM Server is also HP-UX 11i. Robin Sharpe Berlex Labs
Re: strange problem - missing file on server
Nope... checked that. I was thinking it might have something to do with the permissions, but our backups run as root, and other files are backed up. -RS Andrew Raibeck [EMAIL PROTECTED] COM To: ADSM-L@VM.MARIST.EDU Sent by: ADSM: cc: Dist StorSubject: Manager Re: strange problem - missing file on server [EMAIL PROTECTED] T.EDU 06/30/2005 11:29 AM Please respond to ADSM: Dist Stor Manager Is it possible that the file is being inadvertently excluded via some include/exclude pattern? Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-06-30 08:10:46: Hi All, Has anyone run into this situation before? I was asked to restore a single file on a Unix (HP-UX 11i) server. So I fired up the web client, looked in the specified directory, no file, not active or inactive. The file does exist on the client, with a date stamp of Jan 28 2003, and permissions rwxr-x---. There are other files in the directory with same permissions and old (even older) date stamps, that do exist on the server. We keep logs of our daily backups, so I cthe most recent one, which was run earlier the same day (yesterday)... the directory was backed up, but the file was not. The file's name is bkupscripts... nothing weird, no special characters or backspaces in the name a 'find . -name bkupscripts' command does find the file, as does a 'ls -l' command. Client level is 5.2.0.0, Server is 5.2.4.2 (although client reports it as 5.2.3.2... don't know why) TSM Server is also HP-UX 11i. Robin Sharpe Berlex Labs
Re: strange problem - missing file on server
Thanks Richard, I did use the dsmc yesterday also, and it was not there. But I just check again and it did get backed up last night. I'm now thinking that someone moved this file into this directory yesterday, but it retained its ld date stamp... I'll have to ask where it came from. Regards Robin |-+--- | | Richard Sims| | | [EMAIL PROTECTED]| | | | | | 06/30/2005 11:34| | | AM | | | | |-+--- --| | | | | |To: Robin Sharpe [EMAIL PROTECTED] | |cc: | |Subject: | |Re: strange problem - missing file on server | --| Robin - First, do 'dsmc q inclexcl' to see if the file is excluded from play. Second: Don't trust the product GUIs - there's always something wrong with them. The GUIs are provided for novice users. Experienced systems people avoid them. Do 'dsmc q backup -inactive /PathTo/bkupscripts' to really verify. Richard Sims On Jun 30, 2005, at 11:10 AM, Robin Sharpe wrote: Hi All, Has anyone run into this situation before? I was asked to restore a single file on a Unix (HP-UX 11i) server. So I fired up the web client, looked in the specified directory, no file, not active or inactive. The file does exist on the client, with a date stamp of Jan 28 2003, and permissions rwxr-x---. There are other files in the directory with same permissions and old (even older) date stamps, that do exist on the server. We keep logs of our daily backups, so I cthe most recent one, which was run earlier the same day (yesterday)... the directory was backed up, but the file was not. The file's name is bkupscripts... nothing weird, no special characters or backspaces in the name a 'find . -name bkupscripts' command does find the file, as does a 'ls -l' command. Client level is 5.2.0.0, Server is 5.2.4.2 (although client reports it as 5.2.3.2... don't know why) TSM Server is also HP-UX 11i. Robin Sharpe Berlex Labs
Re: Server Locked Up
I've seen similar symptoms when a process needs a scratch tape... TSM will reserve a tape drive, look for a scratch, kill the process, then retry the process and go through the same cycle. Usually with Reclaims, which retry every minute. Also, if there are more than one processes looking for scratches (and none are available), everything else seems to stop until the requests get satisfied or the processes are canceled. SO, I'd look to see if you have any scratches in the library. Robin Sharpe Berlex Labs Warren, Matthew (Retail) [EMAIL PROTECTED] To: ADSM-L@VM.MARIST.EDU RGEN.CO.UK cc: Sent by: ADSM: Dist Subject: Stor ManagerRe: Server Locked Up [EMAIL PROTECTED] U 06/17/2005 10:14 AM Please respond to ADSM: Dist Stor Manager You may try looking through the activity log. Also, for example, if you are backing up storage pools used to hold image backups, because the process does not update it's 'bytecount' untill it finishes a file, and image backup files can be huge, you may find everything is actually working fine. How long has it appeared the server has 'locked up', and is it just the backup stgpools that appear frozen? Matt. Computer says 'No'.. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Debbie Bassler Sent: Friday, June 17, 2005 1:57 PM To: ADSM-L@VM.MARIST.EDU Subject: Server Locked Up It seems our server has locked up. We are copying our storage pools to our copypool and nothing seems to be moving, no counts are increasing. Our server level is 5.1.1. We have 10 tape drives, all online. 8 are in use and 2 are reserved. Has anyone had this problem before? The dsmserv.err log doesn't tell me anything, is there somewhere else I can look? Thanks for any help, ___ Disclaimer Notice __ This message and any attachments are confidential and should only be read by those to whom they are addressed. If you are not the intended recipient, please contact us, delete the message from your computer and destroy any copies. Any distribution or copying without our prior permission is prohibited. Internet communications are not always secure and therefore Powergen Retail Limited does not accept legal responsibility for this message. The recipient is responsible for verifying its authenticity before acting on the contents. Any views or opinions presented are solely those of the author and do not necessarily represent those of Powergen Retail Limited. Powergen Retail Ltd is authorised and regulated by the Financial Services Authority for the sale and service of general insurance products. Registered addresses: Powergen Retail Limited, Westwood Way, Westwood Business Park, Coventry, CV4 8LG. Registered in England and Wales No: 3407430 Telephone +44 (0) 2476 42 4000 Fax +44 (0) 2476 42 5432
Re: Conflicting slot counts
Don't know if this applies to the IBM library, but when it happens on our STK L700, it indicates there are tapes in the library that TSM has lost track of. They need to be checked in... so we do two commands: checkin libv libname search=y checkl=b stat=scr checkin libv libname search=y checkl=b stat=pri These commands are quick and safe... if there are no tapes to checkin, nothing will happen (and you will still see a discrepancy). You do the stat=scr first because TSM will not allow private tapes to be checked in as scratches, but it will allow scratch tapes to be checked in as private (if that happens, you will see private tapes that are not in any storage pool... we call these orphans). Anyway, the audit does not remedy this situation. Robin Sharpe Berlex Labs Zoltan Forray/AC/VCU [EMAIL PROTECTED] To: ADSM-L@VM.MARIST.EDU cc: Sent by: ADSM: Subject: Dist StorConflicting slot counts Manager [EMAIL PROTECTED] T.EDU 06/08/2005 06:57 PM Please respond to ADSM: Dist Stor Manager Why is my AIX TSM server lying about the slot count in my 3583 Library ? When I check the 3583 via the panel, it says there are 3-free slots. TSM refuses to check any more tapes into this library, saying it is full ! Yes, I have done an audit of the library (barcodes)...numerous times !
Re: very very big DB!!!!
FYI... my database is 300GB! (Yes, I am planning to split it): tsm: TSM_WAYNEq db f=d Available Space (MB): 307,852 Assigned Capacity (MB): 307,852 Maximum Extension (MB): 0 Maximum Reduction (MB): 18,848 Page Size (bytes): 4,096 Total Usable Pages: 78,810,112 Used Pages: 73,989,129 Pct Util: 93.9 Max. Pct Util: 93.9 Physical Volumes: 23 Buffer Pool Pages: 300,000 Total Buffer Requests: 796,977,754 Cache Hit Pct.: 95.48 Cache Wait Pct.: 0.00 Backup in Progress?: No Type of Backup In Progress: Incrementals Since Last Full: 0 Changed Since Last Backup (MB): 2,874.37 Percentage Changed: 0.99 Last Complete Backup Date/Time: 06/08/05 13:08:57 tsm: TSM_WAYNE Robin Sharpe Berlex Labs Douglas Currell [EMAIL PROTECTED] .COMTo: ADSM-L@VM.MARIST.EDU Sent by: ADSM: cc: Dist StorSubject: Manager Re: very very big DB [EMAIL PROTECTED] T.EDU 06/09/2005 10:23 AM Please respond to ADSM: Dist Stor Manager 30 GB is not outrageously large. I've worked with one that was almost twice that size. You can unload and load the database or perhaps do an audit on it. There's an entire section called utilities in the Admin Reference. Be forewarned that an audit could take days if not weeks...!!! --- Mario Di Pede [EMAIL PROTECTED] wrote: Hi I have a TSM DB of 30 Gb - there are a utilities for cleaning its? thank's all __ Do You Yahoo!? Tired of spam? Yahoo! Mail has the best spam protection around http://mail.yahoo.com
Re: Data for 600,000 Time Warner employees MIA
Yes, indeed. If those were TSM backup tapes, how much concern would it be? We know that without the TSM server, it would be impossible to identify what files were on the tapes. But would it still be possible to dump the tape and recognize name and SSN data? How difficult (or trivial) would it be to dump a TSM tape in raw format and identify sensitive information? This would be an interesting experiment, I think... Robin Sharpe Berlex Labs |-+ | | Joe Crnjanski| | | [EMAIL PROTECTED]| | | ETWORK.COM | | | Sent by: ADSM: Dist | | | Stor Manager| | | ADSM-L@VM.MARIST.EDU| | | | | || | || | | 05/02/2005 05:58 PM | | | Please respond to| | | ADSM: Dist Stor | | | Manager | | || |-+ | | | | | |To: ADSM-L@VM.MARIST.EDU | |cc: | |Subject: | |Data for 600,000 Time Warner employees MIA | | ! Interesting story !!! Joe Crnjanski Infinity Network Solutions Inc. Phone: 416-235-0931 x26 Fax: 416-235-0265 Web: www.infinitynetwork.com From: ZDNet Must-Read News Alerts [mailto:[EMAIL PROTECTED] Sent: Monday, May 02, 2005 5:01 PM To: Joe Crnjanski Subject: Data for 600,000 Time Warner employees MIA | ZDNet Must-Read News Alert Trouble viewing this mail? Read it online. http://nl.com.com/view_online_newsletter.jsp?list_id=e589 Must-Read News Alert IT News Happening Now http://i.i.com.com/cnwk.1d/i/z/0604/zd_toptab_r.gif Data for 600,000 Time Warner employees MIA http://ct.zdnet.com.com/clicks?c=152808-398764brand=zdnetds=5 Personal information for 600,000 current and former Time Warner employees has been lost, the company announced on Monday, potentially setting the stage for one of the largest cases yet of identity theft. Read the complete story. http://ct.zdnet.com.com/clicks?c=152809-398764brand=zdnetds=5 Is this story important? http://nl.com.com/poll.ss?pollId=28answerId=91id=398764list_id=e589; brand=zdnet Very Important http://nl.com.com/poll.ss?pollId=28answerId=91id=398764list_id=e589; brand=zdnet http://nl.com.com/poll.ss?pollId=28answerId=92id=398764list_id=e589; brand=zdnet Important http://nl.com.com/poll.ss?pollId=28answerId=92id=398764list_id=e589; brand=zdnet http://nl.com.com/poll.ss?pollId=28answerId=93id=398764list_id=e589; brand=zdnet Not Important http://nl.com.com/poll.ss?pollId=28answerId=93id=398764list_id=e589; brand=zdnet TalkBack http://ct.zdnet.com.com/clicks?c=152810-398764brand=zdnetds=5 Is Time Warner at fault for the lost data? http://ct.zdnet.com.com/clicks?c=152811-398764brand=zdnetds=5 (voice your opinion) Customize Your Alerts: Security threats http://ct.zdnet.com.com/clicks?c=152812-398764brand=zdnetds=5 | Data backup, disaster recovery http://ct.zdnet.com.com/clicks?c=152813-398764brand=zdnetds=5 | Most Popular Alerts by Topic: http://i.i.com.com/cnwk.1d/i/zd/fd/2003/08/bullet_yellow.gif Viruses and Worms Alerts http://ct.zdnet.com.com/clicks?c=152814-398764brand=zdnetds=5 http://i.i.com.com/cnwk.1d/i/zd/fd/2003/08/bullet_yellow.gif Hacking Alerts http://ct.zdnet.com.com/clicks?c=152815-398764brand=zdnetds=5 http://i.i.com.com/cnwk.1d/i/zd/fd/2003/08/bullet_yellow.gif Windows XP Alerts http://ct.zdnet.com.com/clicks?c=152816-398764brand=zdnetds=5 http://i.i.com.com/cnwk.1d/i/zd/fd/2003/08/bullet_yellow.gif Spam and phishing Alerts http://ct.zdnet.com.com/clicks?c=152817-398764brand=zdnetds=5 http://i.i.com.com/cnwk.1d/i/zd/fd/2003/08/bullet_yellow.gif Firewalls Alerts http://ct.zdnet.com.com/clicks?c=152818-398764brand=zdnetds=5 http://i.i.com.com/cnwk.1d/i/zd/fd/2003/08/bullet_yellow.gif Spyware/adware Alerts http://ct.zdnet.com.com/clicks?c=152819-398764brand=zdnetds=5 Most Popular Alerts by Keyword: http://i.i.com.com/cnwk
Splitting TSM ...
Dear Colleagues, Well, I've decided it's just about time to bite the bullet and split my 260GB TSM database into two or more. My plan is to have a TSM_DR (disaster recovery), and maybe two others (TSM_A and TSM_B), which can have their load split by client platform or TSM client type or some other scheme to attempt to equalize the DB size. The TSM_DR will only contain those clients that we have designated to be reovered in a site disaster... this will reduce the time required to restore the database at the DR site. I plan to share our STK L700, with the TSM_DR acting as library server. I've already tested this and I'm satisfied that it will work. My question for you all is, how to move the DR storage pools and volumes from the existing TSM to TSM_DR. We've come up with two possibilities: - Export/Import... this is the obvious approach, and probably what Tivoli would recommend, but it will take a while. - Duplicate the existing database, and then delete what does not belong in each; then shrink them down to appropriate size. I think this could be done in a day. Does anyone see any reason why the second approach would not work? Is there another method that I am missing? Thanks in advance for any comments... Robin Sharpe Berlex Labs
ANS1329S (RC29) Server out of data storage space..... But not really!
Has anyone encountered the subject message? We saw it today in a Oracle RMAN backup... what I think happened is, at the time the backup script started, there was not enough space in the ORACLE_DISK pool, nor in the ORACLE_LTO2 pool which is next in the heirarchy. Other sessions of the same job were still working, and we added scratches and ran a migration while this backup job was still running. The script terminated much later, I think when the other sessions finished... but why couldn't it recover and use the newly available disk space or scratch tapes? There isn't a whole lot of detail in the TDP manual about how server space gets reserved and allocated... but it seems from experience that TDP looks for enough space to hold the entire full backup. Can anyone confirm this? TIA Robin Sharpe Berlex Labs
Re: ANS1329S (RC29) Server out of data storage space..... But not really!
Yes, I checked all the pertinent things... at the time the message was issued, the disk pool had 85GB available, the tape pool had over 600GB available (on FILLING volumes), and there were 4 scratches up for grabs. The tape pool has maxscratch=999, and there are only 55 volumes currently in the pool. Maximum size threshold for the disk pool is No Limit. The database being backed up is a total of about 89GB, but there are several table spaces, and I doubt that the object being backed up was 85GB. Other objects from the same backup that succeeded earlier were around 18GB. Couple things I don't know: - how to determine the size of the current object (only see that after it is done) - how does RMAN determine the object sizes? - what size does TDP/API reserve? size of entire data base? (makes sense) - why doesn't it move on to the tape pool? (my guess is, once it tries to allocate to disk, it stays on disk, unlike the BA client) Anyone know of any tech resources (presentations, white papers, etc.) that explain the process TDP/API use to request space? (BTW, my TSM server is 5.2.3.2, TDP 5.2, API 5.1.5, BA client 5.1.1) Robin Sharpe Berlex Labs Ben Bullock [EMAIL PROTECTED] .COMTo: ADSM-L@VM.MARIST.EDU Sent by: ADSM: cc: Dist StorSubject: Manager Re: ANS1329S (RC29) Server out of data storage space. But not really! [EMAIL PROTECTED] T.EDU 02/09/2005 12:10 PM Please respond to ADSM: Dist Stor Manager Did you check the setting of the maxscratch on the tapepool, and then check to see how many tapes are in use? Ben -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Robin Sharpe Sent: Wednesday, February 09, 2005 10:04 AM To: ADSM-L@VM.MARIST.EDU Subject: ANS1329S (RC29) Server out of data storage space. But not really! Has anyone encountered the subject message? We saw it today in a Oracle RMAN backup... what I think happened is, at the time the backup script started, there was not enough space in the ORACLE_DISK pool, nor in the ORACLE_LTO2 pool which is next in the heirarchy. Other sessions of the same job were still working, and we added scratches and ran a migration while this backup job was still running. The script terminated much later, I think when the other sessions finished... but why couldn't it recover and use the newly available disk space or scratch tapes? There isn't a whole lot of detail in the TDP manual about how server space gets reserved and allocated... but it seems from experience that TDP looks for enough space to hold the entire full backup. Can anyone confirm this? TIA Robin Sharpe Berlex Labs
TSM DB -- How big is too big?
Fellow TSM experts, I know this question has no definitive answer, but I have to ask because of our situation. My TSM database is currently about 220GB, 98% utilized and growing steadily. It hasn't posed a serious problem, other than having to throw disk at it. Backup takes 1.5 - 2 hours to LTO2 tape. I also have an overflowed STK L700 library... 14 LTO2 drives, 6 DLT drives, 618 slots about 450 carts on a rack in the computer room I'm considering a couple of options... 1) Freeze current database (shut it down) and start over with a new one. May need to bring the old one up for occasional restore... lots of unanswered questions here, like how will this work with RMAN - TDP Oracle? 2) Split into two or more TSM's -- maybe one for Disaster Recovery (i.e. hot site) machines, and two or three more split by platform or application. Both options will require sharing our library (until we get an additional one), and I'm not sure that can be done by multiple TSMs on the same host machine. Both options create a much more complex TSM environment to manage... right now everything is nice and simple - one TSM, one library. So, I'm open to suggestions comments is a 220GB database justification to create a more complex environment, and is it feasible on one host? (Host is an HP rp7410 running HP-UX 11i, and could be partitioned if needed). Thanks Robin Sharpe Berlex Labs
Re: TSM DB -- How big is too big?
How long does your expiration run and how have you managed it? Well, that's an interesting story... we are not doing any expirations right now due to reasons beyond my control. When we were running expirations, it ran about an hour but the database was only about 180GB then. Robin Sharpe Berlex Labs
Re: Tape Questions
That's way off. There are so many tape technologies available, a single rule of thumb is meaningless. Also it is very dependent on how you configure your TSM... how many storage pools, do you make offsite copies for all data, are you collocating, etc. But, here's a real-world example (mine): We use a mixture of DLT-IV (native capacity 40GB per cartridge) and LTO2 (native capacity 200GB per cartridge). We have hardware compression on, and client compression off. We use collocation for business-critical systems. We are backing up about 30 Unix servers and over 200 Windows servers. (current license in use count is 296) We do daily backups for virtually all clients. We have a total of 500,019,525 Megabytes backed up... that's about 500 Terabytes. We have 1,065 cartridges onsite... 129 DLT's, and the rest LTO2 We have 1,439 cartridges offsite (DRM managed)... 737 DLT, the rest LTO2, and most of these are far less than 50% full. So, 500 Terabytes on 2500 tapes, and with some reclamation we could bring that down significantly. I consider our backup data to be average, or typical, if there is such a thing... almost half is email (Lotus Notes), a quarter is Oracle database, and a quarter is flat files (Windows and Unix). As for growth, we backup almost a terabyte a day, and our current LTO2 tape consumption is about 140 cartridges per month... but that includes both onsite and offsite copies... so 60 TB per month on 140 tapes. This is a high figure because we are temporarily (I hope) not allowed to delete anything... so no files are expiring. We don't use DLT anymore, just waiting for the old ones to expire. Hope this helps... Happy Holidays! Robin Sharpe Berlex Labs |-+--- | | Lepre, James | | | [EMAIL PROTECTED]| | | | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 12/22/2004 09:45| | | AM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- | | | | | |To: ADSM-L@VM.MARIST.EDU | |cc: | |Subject: | |Re: Tape Questions | | Hello All, I have a question is there a rule of thumb for how many tapes will be used, when using TSM. I heard that for every Terabyte that is backed up, 1000 tapes could be used. I was just wondering, if anybody knew something different. I keep 2 versions for 7 days Thanks James
Re: Need to save permanent copy of all files currently being stored
Is all that really necessary? How about creating a new permanent retention domain, copy all relevant policy sets, management classes, copygroups, etc. to the new domain, but change all retentions to NOLIMIT. Then move the affected client to the new domain. Next incremental should rebind all existing data to the new NOLIMIT management classes. I recently had to do something similar, but for ALL clients. That's right, ALL we are currently retaining all data permanently, hopefully only temporarily. One interesting benefit of this excercise is that I now know from real experience what the media cost is in our environment... and it approximately doubled our LTO2 media consumption. Robin Sharpe Prather, Wanda [EMAIL PROTECTED] HUAPL.EDU To: [EMAIL PROTECTED] Sent by: ADSM: cc: Dist StorSubject: Manager Re: Need to save permanent copy of all files currently being stored [EMAIL PROTECTED] T.EDU 12/01/2004 11:13 AM Please respond to ADSM: Dist Stor Manager Mark's right, there is no GOOD way to do this. You can't create an archive with data that is already on TSM tapes, except as a backupset, which can be a pain to track and can't be browsed. I've done this by using EXPORT along with some management class games (Mark, if you can find a hole in this scheme, PLEASE let me know!): * Export THISNODEs filespace that contains the desired directory; make sure you specify ALL to get the inactive data * Rename THISNODE to TEMPCLIENT * Import the filespace * Now you have a second copy of that filespace under the original name THISNODE * Rename the imported THISNODE to something new, like THISNODE-PERM * Rename TEMPCLIENT back to THISNODE. THISNODE is now back to normal. * Copy the POLICY DOMAIN to a new domain called DOMAIN_PERM. Use COPY, not DEFINE to make sure the management classes have the same names. * Change THISNODE-PERM to the new domain (update node THISNODE-PERM DOMAIN=DOMAIN_PERM; no data actually moves). * Change the backup copy groups in DOMAIN_PERM so that nothing ever expires. * Remember to activate the policy set to make the changes effective It's a royal pain in the patootie, because of the time it takes to EXPORT and IMPORT, and like Mark said, you have to take the WHOLE filespace. It's something you don't want to do if the node owns terabytes of data. BUT, it leaves a perm copy of THISNODEs filespace that can be browsed with the GUI, restored, etc., by pointing the client to THISNODE-PERM. And the data is on normal storage pool tapes, so it is managed with your other copy pool tapes. And you don't have to mess with tracking the extra EXPORT or BACKUPSET tapes, or wondering what's on them. Hope that helps. Wanda Prather I/O, I/O, It's all about I/O -(me) -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Stapleton, Mark Sent: Wednesday, December 01, 2004 10:51 AM To: [EMAIL PROTECTED] Subject: Re: Need to save permanent copy of all files currently being stored From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Kevin Kinder I have been ordered to create a permanent copy of one directory (including all files and subdirectories that I currently have stored on TSM for a particular client. These files currently have a 30 day retention, and many of them change daily, so I have 30 copies of a number of these files. What is the best way (any way) to move everything on the backed up directory (both active and inactive files and folders) over to an archive that has no limits on retention? We currently do not utilize archives, only backups, so this is my first experience in this area. I've read the client and admin reference, but I don't see anything that helps me achieve this. Thanks for your help! Well, you can't archive data from a backup. However, you can do one of two things: 1. Create a backupset 2. Export the node Both have drawbacks. They can both be performed at the filesystem level, but not the directory level. An export does not have an expiration date, and a backupset's expiration can be set to NOLIMIT. If you do an export, you probably want to use the FILEDATA=ALLACTIVE flag. You cannot browse an export in order to perform an import, and an import will overwrite all data concerning the node. A backupset's contents can be examined by using the QUERY BACKUPSETCONTENTS command, but you cannot browse in order to select individual files for restore; the only way to bring back an individual file is to know the file's nane and the directory it is located in. The most pertinent question is: why is a permanent copy needed
Re: Infrastructure design questions -- I need input please
We too, are considering expanding into a site about 8 miles away... currently there are servers there that backup across the WAN (about 200GB per night) and it is not a problem. But, for DR purposes, we are considering another TSM server and library at the new location we will probably split the backup workload. In the event of a disaster, we might restore the destroyed TSM DB alongside the incumbent, and do library sharing. Kevin, one thing I'd reconsider is not having a second tape copy we have had many files become unavailable on the primary media (LTO2) for one reason or another... if not for the copy, we would have not been able to recover those files. Robin Sharpe Berlex Labs |-+--- | | Prather, Wanda| | | [EMAIL PROTECTED]| | | HUAPL.EDU | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 07/28/2004 10:28| | | AM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- ---| | | | | |To: [EMAIL PROTECTED] | |cc: | |Subject: | |Re: Infrastructure design questions -- I need input please | ---| I would recommend using the second server only in the event of a disaster. Since you are connected by fibre, the primary server can send the data directly to the tape drives in the library at fibre speeds. You don't want to try and make the 2 servers talk to each other via server-to-server communications, 'cause that will just slow you down to TCP/IP speeds. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Thach, Kevin G Sent: Wednesday, July 28, 2004 9:39 AM To: [EMAIL PROTECTED] Subject: Infrastructure design questions -- I need input please My organization is developing a DR hotsite at one of our other facilities across town, and we are considering making some radical changes to our TSM environment. I know there are several folks on this list that are heavy into TSM design and I could use all the input I can get. Our current environment consists of the following: * TSM server running 5.1.7.3 on AIX. The server is a 6-processor 6H1 w/ 8GB RAM and four 2Gb HBAs. * Approximately 350 clients, and backup 1.5 TB nightly * We use a SAN-attached 3584 with 12 LTO-1 tape drives. 60-day retention policy for everything, so we are maintaining ~90 TB in our local and offsite (copypool) tape pools. * Disk storage pools, DB, Log, are all on SAN-attached IBM Shark disk Our objective is to take advantage of the hotsite not only to improve our DR methods, but to improve TSM restore times. This is what we're considering: * Purchasing approximately 120-140 TB worth of SATA disk, which will live at our current site. All backup data will be retained on disk which should improve restore performance. * Move the tape library to the hotsite, and install a second TSM server there as well. We would no longer create two tape copies of our data, but we would create a single tape copy across town. * The two sites will be connected by dark fiber, so the speed at which we can deliver the data to the 3584 should not be a problem. Is anyone doing something similar to this? Are there any major flaws that I'm not considering? Any advice and input is appreciated. Also, I realize I need to go back and brush up on my TSM manuals, but since I don't run a two-TSM server environment, I have forgotten exactly how that will work in the kind of situation I describe. Would I only use the secondary server in the event of a disaster on the primary? Or would the secondary server at the hotsite manage the library? Etc? If someone can point me in the right direction on that aspect, I'd appreciate it. Thanks
Re: LTO Tape Question
Kelli, If David's suggestion (checkout libv libname 05 checkl=n rem=n) does not help, I would suggest looking at the L700 web page (hope you have it installed!) I can tell you what , if anything is in slot XXX. Robin Sharpe Berlex Labs |-+--- | | Jones, Kelli | | | [EMAIL PROTECTED]| | | IELD.GOV | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 03/10/04 11:13 | | | AM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- --| | | | | |To: [EMAIL PROTECTED] | |cc: | |Subject: | |LTO Tape Question | --| TSM 5.1.7.0 AIX 5.0 STK L700 Library IBM LTO Drives I have a volume (05) that is giving me a fit! Somehow, during library servicing ( I think), volume was removed from library. The volume contains a small amount of data, but contains data none the less. There are no read or write errors. However, when I attempt to check it back in (I have used the search=bulk command, I have named the volume, etc, etc) it fails with an ANR8942E. The message sometimes states: 'Cannot move volume NOT KNOWN from slot 10 (cap) to clot XXX' - (the slot # varies) but it also sometimes says: 'Cannot move volume 05 from slot 10 to slot XXX'. I have run audits and I have tried labeling it (overwrite=no), I have tried moving the data (which fails because it needs the volume to be checked in and it can't be checked in!) Of course, auditing the volume doesn't work for the same reason. I am probably missing something very simple I'm sure but I just can't put my finger on it...any help would be appreciated. Thanks! Kelli Chesterfield County [EMAIL PROTECTED] 804-748-1951
Re: RAW vs. JFS question
Hi all... Wow, this is a controversial subject. We run our TSM 5.2 server on an HP rp7410 server. I just checked the Platform Specific Recommendations for HP-UX, and they recommend allocating disk storage pools on Raw partition, but allocating the DB and log on filesystems! I'm assuming that when they say raw partitions, they really mean raw logical volumes.;) Here is a link to the page: http://publib.boulder.ibm.com/tividd/td/TSMM/SC32-9101-01/en_US/HTML/SC32-9101-01.htm#_Toc58484204 Regards Robin Sharpe Berlex Labs |-+--- | | Zoltan | | | Forray/AC/VCU | | | [EMAIL PROTECTED]| | | | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 01/23/04 08:21 | | | AM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- -| | | | | |To: [EMAIL PROTECTED] | |cc: | |Subject: | |Re: RAW vs. JFS question | -| Well then, IBM needs to revisit their own technical documents. From a recent Technical Exchange series on TSM Performance Tuning: OS Platform Specific Recommendations - UNIX Use raw logical volumes for best TSM server performance AIX: mklv -ae -t tsmdb -y db1 volgrp1 64 hdisk4 mklv -ae -t tsmlg -y lg1 volgrp1 64 hdisk5 TSM: define dbvol /dev/rdb1 define logvol /dev/rlg1 James Hunt [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 01/22/2004 03:01 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Subject:Re: RAW vs. JFS question There are not any performance improvements when the TSM Recovery Log or TSM Data Base is put on RAW volumes. It would be much easier to manage if the environment is all JFS. There may also be a need in the future to run multiple instances of the TSM server code on the same box. Which RAW could cause problems. Also there are some concerns when working with RAW volumes. Do not use AIX to mirror Raw Logical Volume with TSM. Instead use TSM mirroring facilities. AIX puts information in a control block on disk which will overwrite TSM's control information. Do not change the size of RLV's using SMIT. Please refer to 'TSM for AIX Administrators manual for EXTend DB and EXTend LOG commands. The TSM Administrator must be careful not to start multiple instances of TSM Server code which may use the same RLV's. With AIX there is no locking when using RLV. TSM attempts to fix this by implementing locks using files in the /tmp directory. These locks may be deleted if programs are run which clean up the files in the /tmp directory. Jim Hunt ___ GlassHouse Technologies, Inc. [EMAIL PROTECTED] -Original Message- From: Paul Zarnowski [mailto:[EMAIL PROTECTED] Sent: Thursday, January 22, 2004 1:36 PM To: [EMAIL PROTECTED] Subject: Re: RAW vs. JFS question At 06:24 PM 1/22/2004 +0100, you wrote: On Thursday 22 January 2004 15:38, Bill Boyer wrote: The new TSM runing guide recommends using RAW volumes for performance reasons. I think IBM's recommendation is raw for DB and LOG, and JFS for stgpool volumes (because of potential readahead benefits in JFS). And what about the DirectIO and AsyncIO? If you use DirectIO, don't you bypass the file system cache? DirectIO is only usable on AIX for LVs under 2G. That's pretty restrictive. Stef ..Paul -- Paul Zarnowski Ph: 607-255-4757 719 Rhodes Hall, Cornell UniversityFx: 607-255-8521 Ithaca, NY 14853
Re: Upgraded to TSM 5.2.1.2
This is a problem that keeps getting fixed and then coming back in almost every release! IMHO, the real problem is the way the STATUS field is formatted in the PROCESSES table. It's formatted to be readable by humans. It's very difficult to parse in scripts because it contains new lines to make it look nice on screen. Some of the information in STATUS is not available anywhere else (Like current input and output volume, waiting for mount of whatever volume, etc.), and some is redundant (files processed and bytes processed)... however I have found that the files and bytes in the STATUS field gets updated before (maybe more frequently than) the FILES_PROCESSED and BYTES_PROCESSED fields. I wish all of the info in the STATUS field could be placed in individual fields in the PROCESSES table, and fix this problem once and for all. It would save us script writers lots of rework time for each new release! Robin Sharpe Berlex Labs |-+--- | | Ben Bullock | | | [EMAIL PROTECTED]| | | .COM | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 12/10/03 10:10 | | | AM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- -| | | | | |To: [EMAIL PROTECTED] | |cc: | |Subject: | |Re: Upgraded to TSM 5.2.1.2 | -| Yes, we upgraded to 5.2.1.3 on AIX and have the odd formatting in the q proc output. No fix that I know of, luckily it is just cosmetic as far as I can tell. Ben -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Kamp, Bruce Sent: Wednesday, December 10, 2003 5:58 AM To: [EMAIL PROTECTED] Subject: Upgraded to TSM 5.2.1.2 I just upgraded to TSM 5.2.1.2 from 5.1.1.6 on AIX 5.1ML5 64bit. One thing I noticed is that when I do a Q PR it is not formated correctly on the screen. Has anybody else seen this is there a fix for it? Also is there anything else I need to watch out for? Thanks, - Bruce Kamp Midrange Systems Analyst II Memorial Healthcare System E-mail: [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] Phone: (954) 987-2020 x4597 Pager: (954) 286-9441 Alphapage: [EMAIL PROTECTED] Fax: (954) 985-1404 -
Upgrade TSM 5.1.7 to 5.2.0 - how long should upgradedb take?
Hi all, Today we started our upgrade of TSM 5.1.7 to 5.2.0. Our TSM server is HP-UX 11i, TSM database is 112GB, about 80% used. After the install of the TSM product, the 'dsmserv -upgradedb' command is run during server reboot (that's how it is done on HP-UX the startup screen says Configuring all unconfigured software filesets). At this point, the system is not responding to pings... we suspect it is performing the upgradedb in single-user mode, or maybe just hasn't started network processes yet. The system has been in this state for over three hours is this normal? I don't recall the upgrade from 4.1 to 5.1 taking this long. Thanks for any tips... Robin Sharpe Berlex Labs
Re: Online DB Reorg
Mine came back as 0.02 does that mean it is not very fragmented? Our database is 112GB, 78.8% used, was new in October 2001 and has never been reorganized. Robin Sharpe Berlex Labs |-+--- | | Prather, Wanda| | | [EMAIL PROTECTED]| | | HUAPL.EDU | | | Sent by: ADSM: | | | Dist Stor | | | Manager| | | [EMAIL PROTECTED]| | | T.EDU | | | | | | | | | 10/20/03 01:03 | | | PM | | | Please respond | | | to ADSM: Dist | | | Stor Manager | | | | |-+--- -| | | | | |To: [EMAIL PROTECTED] | |cc: | |Subject: | |Re: Online DB Reorg | -| That query (taken here from ADSM.QuickFacts) confuses me entirely - can someone please explain? SELECT CAST((100 - ( CAST(MAX_REDUCTION_MB AS FLOAT) * 256 ) / (CAST(USABLE_PAGES AS FLOAT) - CAST(USED_PAGES AS FLOAT) ) * 100) AS DECIMAL(4,2)) AS PERCENT_FRAG FROM DB It finds the number of unused pages (usable_pages - used_pages). Then it takes max-reduction and divides by unusable pages. But, so what? I don't get it. The unusable pages - max_reduction tells you how much of your data base is NOT usable. BUT again, so what? That doesn't say whether you need to do a data base reorg or not, does it? If my max reduction is 8 pages and my unused pages are 10, I've got 2 unusable pages. But if my data base is 1,000,000 pages, that certainly isn't much fragmentation, the way a DB administrator (or space manager) would traditionally see it. Certainly no reason to do a DB reorg. WHY isn't the division done with the total usable pages as the numerator? The data base size has to enter in to the decision to reorg, somehwere. I'm confsed (But then, it's Monday) -Original Message- From: Loon, E.J. van - SPLXM [mailto:[EMAIL PROTECTED] Sent: Monday, October 20, 2003 6:38 AM To: [EMAIL PROTECTED] Subject: Re: Online DB Reorg Hi Guys! The SQL statement can also be found at Richard's quickfacts page: http://people.bu.edu/rbs/ADSM.QuickFacts Kindest regards, Eric van Loon KLM Royal Dutch Airlines -Original Message- From: Remco Post [mailto:[EMAIL PROTECTED] Sent: Monday, October 20, 2003 12:30 To: [EMAIL PROTECTED] Subject: Re: Online DB Reorg On Sat, 18 Oct 2003 14:35:09 -0400 Talafous, John G. [EMAIL PROTECTED] wrote: Remco, Would you be willing to share your SQL query that reports on DB fragmentation? I was allready looking at Eric (he probably saved my thingy somewhere usefull, I just saved it in my sent-mail folder), here it is... select cast((100 - ( cast(MAX_REDUCTION_MB as float) * 256 ) / - (cast(USABLE_PAGES as float) - cast(USED_PAGES as float) ) * 100) as - decimal(4,2)) as percent_frag from db Note that I still think this is one of the more useless queries I've ever build... Thanks to all, John G. Talafous IS Technical Principal The Timken CompanyGlobal Software Support P.O. Box 6927 Data Management 1835 Dueber Ave. S.W. Phone: (330)-471-3390 Canton, Ohio USA 44706-0927 Fax : (330)-471-4034 [EMAIL PROTECTED] http://www.timken.com -- Met vriendelijke groeten, Remco Post SARA - Reken- en Netwerkdiensten http://www.sara.nl High Performance Computing Tel. +31 20 592 8008Fax. +31 20 668 3167 I really didn't foresee the Internet. But then, neither did the computer industry. Not that that tells us very much of course - the computer industry didn't even foresee
Re: High CPU Consumption on Tape Threads in DSMSERV
Bill, We have a similar environment -- TSM 5.1.7.x, HP-UX (11i), DLT8000s and LTO2's. I have not noticed the increase you describe, but we have also moved from an 2-CPU L2000 server to an 8-CPU rp7410. Can you be more specific about how you tracked that info with glance? I'll be happy to run the same tests and post my results. Hoe did you test the stape driver? With TSM 5.1.7.2, or are you comparing to the old config? Or stand-alone drives? Regards Robin Sharpe Berlex Pharmaceuticals |-+-- | | Slaughter, Bill | | | [EMAIL PROTECTED]| | | ERWARE.COM| | | Sent by: ADSM:| | | Dist Stor Manager | | | [EMAIL PROTECTED]| | | DU| | | | | | | | | 10/14/03 05:08 PM | | | Please respond to | | | ADSM: Dist Stor | | | Manager | | | | |-+-- -| | | | | |To: [EMAIL PROTECTED] | |cc: | |Subject: | |High CPU Consumption on Tape Threads in DSMSERV | -| I have just had a call escalated to level 2 and was wondering if anybody else is having a problem and may not know it. I was told I was the first to report it. We have recently upgraded to TSM 5.1.7.2 and noticed there is a higher amount of CPU being consumed by the DSMSERV process. We were upgrading from 4.2.3.2. I have narrowed it down using HP-UX glance to be related to tape drive activity and can see (using glance) that there is a correlation between the number of active tape drives and the amount of CPU being consumed. I further investigated and discovered there are many threads consuming CPU when this happens and that 1 thread per Tape DRIVE is 90-100% CPU. I compared these to the HP-UX stape device driver and it will only use 5 - 7% CPU (And it performs much faster). Misery loves company; I would like to know if anybody else is experiencing any of the same problems with high CPU on tape threads and what flavor of UNIX they are using. Bill My Environment: TSM Server 5.1.7.2 TSM Device Driver 5.1.7.2 and 5.1.6.5 HP V2500 32 CPU - HP-UX 11.0 STK 9710 DLT7000 (8 Drives) STK 9710 DLT7000 (6 Drives) Bill Slaughter Tupperware 407-826-4580
Re: High CPU Consumption on Tape Threads in DSMSERV
Bill, I launched some BACKUP STORAGEPOOLs, and checked glance as you suggested. I have about 12 screens' worth of threads under dsmserv (standard 25-line screen size). I did locate a few threads that seem to correspond to the tapes being used by the backup, and they seem to alternate between IO wait and SYSTM wait (which I would expect for I/O processes). The CPU util for these threads is somewhat higher than the other dsmserv threads, but nowhere near what you are seeing mine are around 5%. These were using my DLT8000 drives maybe there's and issue with the older drives, or with HP-UX 11.0? (We have 11i). I'll keep looking over the next few days. Robin |-+-- | | Slaughter, Bill | | | [EMAIL PROTECTED]| | | ERWARE.COM| | | Sent by: ADSM:| | | Dist Stor Manager | | | [EMAIL PROTECTED]| | | DU| | | | | | | | | 10/15/03 10:42 AM | | | Please respond to | | | ADSM: Dist Stor | | | Manager | | | | |-+-- -| | | | | |To: [EMAIL PROTECTED] | |cc: | |Subject: | |Re: High CPU Consumption on Tape Threads in DSMSERV | -| Robin, I selected the threads of the DSMSERV process with glance using option G. All threads were listed (3 pages). If you use the + and go to the last page you may be able to see some waiting on I/O. These are more than likely the Tape threads. When active doing MIGRATION or BACKUP STORAGE POOL these consume high amounts of CPU, SPACE RECLAMATION and MOVE DATA will also but it depends on the density of the data on the tape. I was able to compare a couple of ways, I regressed the driver back to stape for all my tape drives except 1 and can show the high CPU. I also still have a 4.2.3.2 system around and a test system I can configure and play with. Bill Slaughter Tupperware 407-826-4580
Re: How to read LTO cartridge memory (was Re: Media Fault)
Wouldn't that require special hardware? It's radio, isn't it? Robin Sharpe Berlex Labs Anthonijsz, Marcel M SITI-ITDGE13 To: [EMAIL PROTECTED] Marcel.Anthonijsz cc: @SHELL.COMSubject: Sent by: ADSM:How to read LTO cartridge memory (was Re: Media Fault) Dist Stor Manager [EMAIL PROTECTED] EDU 08/26/03 07:14 AM Please respond to ADSM: Dist Stor Manager Hi *SM'ers, Does anybody know how to read the LTO Cartridge memory from the LTO cartridge? The LTO specs/brochure show an expected life cycle of about 1 million mounts and recommends replacement after about 5000 loads. We want to know how close we are to this figure. TSM forgets about mounts as soon as a volume gets scratched... Now did somebody perform the exercise Richards Sims describes below? If not... I see an opportunity here I never did SCSI programming, so there must a first time for everything :-/ Thanks! Marcel Anthonijsz Central Data Storage Manager (a.k.a. storman) Shell Information Technology International B.V. PO Box 1027, 2260 BA Leidschendam, The Netherlands Email: Marcel.Anthonijsz.-at-.shell.com Internet: http://www.shell.com Date: Jul 01, 09:42 From: Richard Sims [EMAIL PROTECTED] Question is: Do you possibly know any software capable of extracting info from LTO CM?? (I mean of course a program that can be run against a suspected cartridge) Wieslaw Now, you know you weren't supposed to ask that question... :-) My research indicates that vendors don't consider that customers should need to access the Medium Auxiliary Memory (MAM) - the industry generic name for an in-cartridge non-volatile memory chip which tracks usage and other info. The manual IBM TotalStorage LTO Ultrium Tape Drive - SCSI Reference (GA32-4050) fully describes their MAM and how to read and write it via SCSI commands. The device driver programming manual (in this case, IBM Ultrium Device Drivers - Programming Reference (GC35-0483)) provides many ioctl functions which make it easier for a programmer to invoke what resolve to SCSI commands; but in this case I see no ready operation for getting MAM data. Those ioctl operations are what the handy-dandy ntutil and tapeutil commands invoke to acquire info, and I see nothing in their doc saying that they can return it (though it might be implicitly returned from other operations). All this is to say that with some SCSI programming, the information could be obtained and presented. We don't have LTO here, so I'm not in a position to try this out. So this remains an exercise for some industrious systems programmer out there having LTO on-site. Richard Sims, Sr. Systems Programmer, Boston University OIT http://people.bu.edu/rbs
Re: how to get multi-session restore to work?
We just did a large restore of an NT box, and got the multi-session restore working. As others have said, make sure the RESOURCEUTILIZATION on the client's dsm.sys is set to more than 1, and the MAXNUMMP for the client's node definition in the server is GREATER or EQUAL to the RESOURCEUTILIZATION (the manual states that if RESOURCEUTILIZATION is greater, you may get errors). Final requirement is that a RESTARTABLE RESTORE is done. The manual (for Windows client) says that choosing any drive or directory will work, but we found that you must select an entire drive to get the multiople sessions. Unix equivalent would be an entire filesystem, although I haven't tested it on Unix yet... it may be a Windows client restriction. Also, the manual says that additional sessions are only started when data resides on a different tape volume, so there *should* not be any contention problems... and we didn't have that issue, with eight sessions running. Good Luck Robin Sharpe Berlex Labs MC Matt Cooper (2838) [EMAIL PROTECTED] To:[EMAIL PROTECTED] TINGS.COM cc: Sent by: ADSM: Subject: Dist Stor Manager how to get multi-session restore to work? [EMAIL PROTECTED] DU 06/03/03 10:16 AM Please respond to ADSM: Dist Stor Manager Hello all, I have an AIX platform that I know had a multi-session backup done. However, the restore seems to be a single thread. What haven't I done? The TSM client is 5.1.0 the server is 5.1.5.4. The restore was initiated with the gui. Matt
Re: Antwort: Volumes Private with NO stgpool last use DATA
Markus, I've seen that also... I think what happens is TSM tries to use a tape that has been deleted as a scratch, but encounters a problem of some kind, so it marks the tape provate to prevent it being used again. If that is the case, you should see a message in the activity log indicating the condition. Do a query of the actlog searching for the tape volser and you should see the message. HTH Robin Sharpe Markus Veit markus.veit@BAY ERBBS.COM To:[EMAIL PROTECTED] Sent by: ADSM: cc: Dist StorSubject: ManagerAntwort: Volumes Private with NO stgpool last use DATA [EMAIL PROTECTED] T.EDU 12/27/02 11:26 AM Please respond to ADSM: Dist Stor Manager Hi, thats the thing, there are no records in the volume history, but still Private with Data Seems like a bug to me, or a feature for tape manufactures :-) Mit freundlichen Grüßen / Best Regards Markus Veit An: [EMAIL PROTECTED] Kopie: Thema: [EMAIL PROTECTED] U Received : 27.12.2002 17:19 Bitte antworten an ADSM: Dist Stor Manager To find out what TSM actually did with that volume, you can use this SQL query: SELECT * FROM VOLHISTORY WHERE VOLUME_NAME='volser' This will give you a history of how TSM used that tape, and the last entry is the final status of the volume. If the tapes are truely not assigned to anything (stgp, dbbackup, backupset, export,...) then just CHECKOUT LIBV with REM=NO and check them back in as STAT=SCR. Bill Boyer DSS, Inc. Some days you're the bug, some days you're the windshield. -- ?? -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED]]On Behalf Of Nelson, Doug Sent: Friday, December 27, 2002 10:58 AM To: [EMAIL PROTECTED] Subject: Database backup tapes will also show up as private, data, with no associated storage pool. Douglas C. Nelson Distributed Computing Consultant Alltel Information Services Chittenden Data Center 2 Burlington Square Burlington, Vt. 05401 802-660-2336 -Original Message- From: Markus Veit [mailto:[EMAIL PROTECTED]] Sent: Friday, December 27, 2002 10:50 AM To: [EMAIL PROTECTED] Subject: Hi TSMers, does anybode have an idea on why volumes in a library have a status of Private with last use Data but do not belong to a storagepool. Volumes with a status of private with no last use are defect, ie can't read the label. volume_name LIBRARY_NAME STATUS LAST_USEACCESS 41L1ADIC100 Private Data 01L1ADIC100 Private Data 68L1ADIC100 Private Data 61L1ADIC100 Private Data 51L1ADIC100 Private Data 47L1ADIC100 Private Data 45L1ADIC100 Private Data 79L1ADIC100 Private
Re: Antwort: Re: Antwort: Volumes Private with NO stgpool last use DATA
In that case, I would try re-labeling with overwrite=yes. If that doesn't work, they may be defective. Were they ever used, or are they brand-new? Robin Markus Veit markus.veit@BAY ERBBS.COM To:[EMAIL PROTECTED] Sent by: ADSM: cc: Dist StorSubject: ManagerAntwort: Re: Antwort: Volumes Private with NO stgpool last use DATA [EMAIL PROTECTED] T.EDU 12/27/02 11:48 AM Please respond to ADSM: Dist Stor Manager Hi Robin, that's true, the message goes something like status of volume blah blah has changed to private to prevent reaccess, but the volume has no last use when you do a q libv volume it is just private, not private and last use data Mit freundlichen Grüßen / Best Regards Markus Veit An: [EMAIL PROTECTED] Kopie: Thema: Re: Antwort: Volumes Private with NO stgpool last use DATA [EMAIL PROTECTED] U Received : 27.12.2002 17:36 Bitte antworten an ADSM: Dist Stor Manager Markus, I've seen that also... I think what happens is TSM tries to use a tape that has been deleted as a scratch, but encounters a problem of some kind, so it marks the tape provate to prevent it being used again. If that is the case, you should see a message in the activity log indicating the condition. Do a query of the actlog searching for the tape volser and you should see the message. HTH Robin Sharpe Markus Veit markus.veit@BAY ERBBS.COM To:[EMAIL PROTECTED] Sent by: ADSM: cc: Dist StorSubject: ManagerAntwort: Volumes Private with NO stgpool last use DATA [EMAIL PROTECTED] T.EDU 12/27/02 11:26 AM Please respond to ADSM: Dist Stor Manager Hi, thats the thing, there are no records in the volume history, but still Private with Data Seems like a bug to me, or a feature for tape manufactures :-) Mit freundlichen Grüßen / Best Regards Markus Veit An: [EMAIL PROTECTED] Kopie: Thema: [EMAIL PROTECTED] U Received : 27.12.2002 17:19 Bitte antworten an ADSM: Dist Stor Manager To find out what TSM actually did with that volume, you can use this SQL query: SELECT * FROM VOLHISTORY WHERE VOLUME_NAME='volser' This will give you a history of how TSM used that tape, and the last entry is the final status of the volume
Re: StorageTek L700E library TSM question
Sounds like there is something wrong with your library. We have an HP 20/700, which is a rebadged STK L700. I'm pretty sure the robot does look at the CAP when we put tapes in without being prompted by TSM. We definitely see tapes in there on the web interface, and checkins work as expected. Do you have one CAP or two? We have two, and it always goes to CAP B (the one on the right) first, for everything. We have another 20/700 at our west coast location, and they have had lots of problems with it... mostly robot/barcode related. Many parts were replaced. They went back-level in firmware to match us. They finally got it stable, but it required many months and intensive involvement of HP top gun CE. We started with TSM 4.1 upgraded to 5.1 on HP-UX server, and the 20/700 worked fone on both (there were new drivers in 5.1 for HP-UX). Robin Sharpe Berlex Labs Spearman, Wayne wmspearman@NOVANTH EALTH.ORG To:[EMAIL PROTECTED] Sent by: ADSM: cc: Dist Stor Manager Subject: [EMAIL PROTECTED]StorageTek L700E library TSM question DU 11/05/02 10:00 AM Please respond to ADSM: Dist Stor Manager Hi All, We have a new STK L700e LTO tape library attached to a new AIX TSM server. We are not running ACLS on the library. I can't get the library to accept tapes via the Cartridge Access Point (CAP). We open the door, place a tape in the slot, close the door, and run the checkin command specifying BULK. The process ends in a failure. The robot never looks at the CAP. When we put the tape in the CAP, the robot never inventories the CAP. I find this odd as our IBM 3494 tape libraries always check the CAP when the door is opened to see what has happened. STK support says their library doesn't do this. STK support says TSM has to tell the robot to go check the CAP. I can't find anything in the TSM books that supports this. The STK web GUI never shows tapes in the CAP either. Can anyone help with this dilemma? Thanks, Wayne Spearman Information Technology - Software Systems Engineer Novant Health - Central Services Charlotte, NC Phone: 704-384-7019 Fax: 704-316-9936 email:[EMAIL PROTECTED] This message and any included attachments are from NOVANT HEALTH INC. and are intended only for the addressee(s). The information contained herein may include trade secrets or privileged or otherwise confidential information. Unauthorized review, forwarding, printing, copying, distributing, or using such information is strictly prohibited and may be unlawful. If you received this message in error, or have reason to believe you are not authorized to receive it, please promptly delete this message and notify the sender by e-mail. Thank you.
Re: Problem tape
We've had problems with mysterious tapes on similar hardware, but HP-UX server. Library is HP 20/700 which is a rebadged L700. Symptoms were TSM being almost hung... could not start new admin sessions, had to kill it restart. we would see two dsmserv processes, which we think is normal, but usually the second one comes and goes so fast that you rarely see it when we were finally able to look at the activity log, we found I/O errors for drives that had sense codes of all , and *unkown* errors... tracing back to mount commands we always foun the same tape mounted in the drive that had the I/O errors removed the tape, and everything was fine. But, we still don't know what was wron with the tapes... don't know if it is a library isuue or a HP-UX issue. Robin Sharpe Berlex Labs Karel Bos Karel.Bos@NU ON.COM To:[EMAIL PROTECTED] Sent by: cc: ADSM: Dist Subject: Stor ManagerRe: Problem tape [EMAIL PROTECTED] RIST.EDU 11/05/02 04:28 AM Please respond to ADSM: Dist Stor Manager We had this on this TSM server version. Please upgrade. -Oorspronkelijk bericht- Van: Chuck Lam [mailto:chuck_lam;YAHOO.COM] Verzonden: maandag 4 november 2002 22:46 Aan: [EMAIL PROTECTED] Onderwerp: Problem tape TSM 4.1.4 running on AIX 4.3.3 Hi, Attached to this server is a StorageTek L700 tape library, after running flawlessly for over a year. I have begun seeing problem tapes popping up here and there. There is this one tape of which I could not mount it for anything, the TSM just gave it up by saying Mount Request denied, but 'q vol' showed nothing was wrong with this tape. Have any ones encountered problem like this, and any suggestions to fix it, other than phsically removing it from within the tape library? Thank you. __ Do you Yahoo!? HotJobs - Search new jobs daily now http://hotjobs.yahoo.com/
Re: volume management in tsm
Justin, You're on the right track. Unfortunately, there is no ideal solution. If you need more slots, you have to move tapes out of the library (use the move media command). TSM will track the locations, and if a reclamation needs a tape, you'll get a mount message in the activity log. Then you have to check the tape back in (stat=private). We are in the same situation we upgraded from an ATL P3000 to a HP 20/700 with 678 slots and now even that is full. I'm considering putting the ATL back into service, but i need to upgrade the drives to DLT8000 first. Regards Robin Sharpe Berlex Labs Justin Bleistein justin.bleistein@SU NGARD.COM To:[EMAIL PROTECTED] Sent by: ADSM: Dist cc: Stor ManagerSubject: [EMAIL PROTECTED]volume management in tsm U 09/26/02 11:51 AM Please respond to ADSM: Dist Stor Manager We are in a bit of a pickle or soon will be anyway We have an IBM/3494 atl in our TSM environment. Which has about 1,000 slots in it. These slots are filling and fast. We're trying to make room for more scratch tapes. Now I've tried everything from move data to weird reclaimation sinareos. And we just can't seem to free up sificiant slots. I had an idea of ejecting all tapes which haven't been written to in a while there just waiting to expire, this way migration or backup storage pool processes won't call for them, and once they're out rack them in the data center and mark there location as: rack. Now that will free up a ton of spots in the atl. The problem is how can TSM manage those tapes even though they are not in the ATL. I can check them in with check label = no, and just mark there location as: rack this way the database will keep updating them. The problem is what if reclaimation runs, now you can't do it because the tapes can't be mounted they're in the rack outside the atl. Unless the robot comes out to get them I don't see how this can work. It seems that trying to come up with a solution to the problem of lack of slots in the atl will just create more of a problem. Any thoughts? Or ideas? on how I can manage these tapes even though they can't be mounted?. P.S. = Yes it is collocated. Let me know if anyone has done this thanks!. --Justin Richard Bleistein
Re: new server migrations
I agree with Nathan... option 1 is not possible. I tried it going from AIX to HP. Option 2 depends on how much data you have. We had about 60TB on our AIX TSM Server in a ATL P3000 (DLT7000 drives). I roughly calculated it would take over six months of tape to tape copies, running 24x7 with optimal throughput, to migrate the environment to the HP. BTW, you must do export/import, because there is no other way to get the database info from AIX to other platform. Even if you could get all the database info migrated, you probably would not be able to read the tapes from the AIX environment on the Sun hardware. It is not a purely hardware issue (obviously you can transfer tar images between platforms), but has something to do with the way TSM records data on the different platforms. That's what Tivoli support told me. My solution to this messy problem was to keep the legacy AIX environment up until all of the data residing there expires (one year, which will be coming up in about a month!). I simply started backing up all clients to the new TSM server, and we handle any restore requests for older data from the AIX server. Obviously you have to have new server and library, but it looks like you have all the pieces as long as you did not have any plans for the legacy hardware. Good Luck! Robin Sharpe Berlex Labs Nathan King [EMAIL PROTECTED] R.COMTo:[EMAIL PROTECTED] Sent by: cc: ADSM: Dist Subject: Stor ManagerRe: new server migrations [EMAIL PROTECTED] RIST.EDU 09/16/02 05:11 PM Please respond to ADSM: Dist Stor Manager Actually No.1 just flat out won't work. You can't backup the TSM DB on one platform and restore to a different platform. No.2 is your only option, in conjunction with import/export -- that's what the tools are for. Don't use server-server. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED]] On Behalf Of Nicola Albrecht Sent: Monday, September 16, 2002 4:49 PM To: [EMAIL PROTECTED] Subject: Re: new server migrations Reiss David IT751 (ext-CDI) wrote: I have several TSM servers that are going to be replaced with new-beefer TSM servers. Right now the TSM servers are running on old RS/6000 boxes running AIX 4.3.3 with Breece Hill Q47 libraries with DLT8000 tapes drives in them. The new TSM Servers are going to be Sun boxes running Solaris 2.8 with ADIC Scalar 1000 libraries with LTO drives. The issue... I want to keep my current TSM data, as much as possible anyway. There seems to be two ways too do this. 1. On the old RS/6000 backup the TSM database to tape. Connect the old Q47 library to the new Sun box, and restore the TSM database. Then, migrate the data via the directly attached SCSI drives from the old library to the new library. Remove old library and beat it to death with nearest sledgehammer. 2. Bring up the new Sun box as a brand new TSM server, define everything from scratch. Then...and here lies the heart of my question, use Server-to-Server stuff to get old data from old Rs/6000 server to new one. I know I can push the data doing this... but could I push it in a logical manner so that it gets associated properly with what will be the node definitions on the new server? I don't like the #1 option because of all the hardware moving things around and such. I would have to visit the places the server will be in. We don't have people at these sites who would be comfortable just switching cables around and stuff. So, sending them a whole new server... have them put it together, and once it is together I can just move on from there and do the data moving myself. Thanks, David N. Reiss TSM Support Engineer [EMAIL PROTECTED] 407-736-3912 David, of the 2 options, number one is the only one that is going to accomplish the association of stored data to node definitions on the new server. The second option (Server-to-Server) stores the data as archive files on the new server and associates these archive file with a node of the type server (old server). Not what it sounds like you want. Another option could be exporting the data, then importing to the new server. Nici Nici Albrecht MDR Consulting Education 1-210-860-4641
Re: Strategies for DR recovery of large clients
Juraj, Your scenario would be helpful to recover a lost filesystem or a lost client in the datacenter, but for Disaster Recovery, we must assume we'll be rebuilding the TSM server plus several critical clients at another site... and that requires tape, unless you can justify spending big bucks to mirror you environment in real time at the recovery center. Robin
Re: Strategies for DR recovery of large clients
Werner, I feel your pain... ;) You have hit most of the major issues of disaster recovery with TSM squarely on the head. We have had similar experience in our testing... 4-6 hours to get the TSM server up, running through loads of tapes (even though storage pool is collocated with only three servers), 48-hour window, etc. We have proved that we can get our three critical clients back within 24 hours, but they are not nearly as big as yours. We use DLT8000 drives. Probably the best way for you to get better restore throughput is to add more drives and do concurrent restores. TSM should only mount the tapes that actually contain the file versions you will restore. The problem is that, even with collocation, after many months of backups on a relatively active system, these files will get scattered across many tapes. Conventional wisdom suggests using collocation by filespace to reduce this effect... and also guarantee that concurrent restores of different file systems will not compete for the same tape volume. But the cost is of course using a lot more tape. Another approach might be to occasionally (every three months maybe) do a full backup (by changing mode to absolute to force even unchanged files to get backed up)... this should effectively defragment the tape pool and put all active versions on one (or a couple) tape. We did this once with an additional machine that we DR'ed and it worked quite well. Some people don't like this concept because it defeats TSM's progressive backup methodology, but I think its an acceptable compromise. As you said, backup sets are not a good option for DR... for one thing, creating the backupset will take as long as restoring the whole system, and will read the same number of tapes. You will suffer this on a regular schedule since you'll have to make new backupsets probably every week or two. Secondly, restoring from backupsets effectively single-threads that client because all of it's data is on one or maybe two tapes. Good luck, and please keep us posted on your results! Robin Sharpe Berlex Labs
Re: BareMetalRestore
I'd be happy if TSM provided a means of managing the media containing my HP Ignite-UX images... currently, that has to be done manually. Robin Sharpe Berlex Labs
Re: Eternal Data retention brainstorming.....
How about a backupset of every node on that day? Of course, it would have to be today I guess... Robin Sharpe Berlex Labs Cook, Dwight E DWIGHT.E.COOK To:[EMAIL PROTECTED] @SAIC.COM cc: Sent by: Subject: ADSM: Dist Re: Eternal Data retention brainstorming. Stor Manager [EMAIL PROTECTED] IST.EDU 08/16/02 07:09 AM Please respond to ADSM: Dist Stor Manager Tell the legal department you want a complete duplicate of you TSM environment !!! People make so many unrealistic request with absolutely no thought as to what they are asking for. The only way to do that is to make a handful of tsm db backups (to protect against loss due to media failure) then take those and ALL private media, box it up and stick it in a hole somewhere. Then if they ever want anything from it, you could do like the last half of your #4 option... Yes, I've had this request before, but it was only for a subset of nodes, I exported them. Took me about a month to do, while I was doing it I turned expiration off, just about ran out of ATL space during the process. I told myself I WOULD NEVER DO THAT AGAIN! If they really want the data, then cost is no object ! and say 400 TB on 3590 K tapes at 3/1 compression is only about 3,333 tapes, at just under $55 ea that is only $183,333.00 CHEAP ! ! ! (oh, that's my supplier's price on K tapes, if that is good or bad ???) Dwight - -Original Message- From: bbullock [mailto:[EMAIL PROTECTED]] Sent: Thursday, August 15, 2002 2:31 PM To: [EMAIL PROTECTED] Subject: Eternal Data retention brainstorming. Folks, I have a theoretical question about retaining TSM data in an unusual way. Let me explain. Lets say legal comes to you and says that we need to keep all TSM data backed up to a certain date, because of some legal investigation (NAFTA, FBI, NSA, MIB, insert your favorite govt. entity here). They want a snapshot saved of the data in TSM on that date. Anybody out there ever encounter that yet? On other backup products that are not as sophisticated as TSM, you just pull the tapes, set them aside and use new tapes. With TSM and it's database, it's not that simple. Pulling the tapes will do nothing, as the data will still expire from the database. The most obvious way to do this would be to: 1. Export the data to tapes store them in a safe location till some day. This looks like the best way on the surface, but with over 400TB of data in our TSM environment, it would take a long time to get done and cost a lot if they could not come up with a list of hosts/filespaces they are interested in. Assuming #1 is unfeasible, I'm exploring other more complex ideas. These are rough and perhaps not thought through all the way, so feel free to pick them apart. 2. Turn off expire inventory until the investigation is complete. This one is really scary as who knows how long an investigation will take, and the TSM databases and tape usage would grow very rapidly. 3. Run some 'as-yet-unknown' expire inventory option that will only expire data backed up ~since~ the date in question. 4. Make a copy of the TSM database and save it. Set the reuse delay on all the storage pools to 999, so that old data on tapes will not be overwritten. In this case, the volume of tapes would still grow (and need to perhaps be stored out side of the tape libraries), but the database would remain stable because data is still expiring on the real TSM database. To restore the data from one of those old tapes would be complex, as I would need to restore the database to a test host, connect it to a drive and pretend to be the real TSM server and restore the older data. 5. Create new domains on the TSM server (duplicates of the current domains). Move all the nodes to the new domains (using the 'update node ... -domain=..' ). Change all the retentions for data in the old domains to never expire. I'm kind of unclear on how the data would react to this. Would it be re-bound to the new management classes in the new domain? If the management classes were called the same, would the data expire anyways? Any other great ideas out there on how to accomplish this? Thanks, Ben
Re: TSM 5.1 HP-UX 11.11 Startup/Shutdown Scripts
Brian, Here's my script, but we run TSM 4.1 (still). Is 5.1 different? You may not want to have user ID and password in a script, which this method requires... I can't think of any other clean way of stopping the server. We had the rm adsmserv.lock in the startup, but removed it because it enables a second startup to run when the server is still up which could be REALLY bad. What's your overall impression of running TSM on HP? We've been doing it for almost a year now (on an L2000), and have some issues with intermittent (but too often) hanging/freezing of TSM... can still get into HP-UX, but dsmadmc is dysfunctional. Robin #!/sbin/sh ## ## File: /sbin/init.d/tsm ## Description: Startup/shutdown script for Tivoli Storage Manager server ## if [ -f /etc/rc.config.d/tsm ];then . /etc/rc.config.d/tsm fi rval=2 case $1 in start_msg) echo Starting Tivoli Storage Manager Server ;; stop_msg) echo Shutting down Tivoli Storage Manager Server ;; start) if [ $RUN_TSM = 1 ];then cd /opt/tivoli/tsm/server/bin #rm ./adsmserv.lock /dev/null 21 #/opt/tivoli/tsm/server/bin/dsmserv -quiet ./dsmserv ./dsmserv.log rval=$? fi ;; stop) dsmadmc -id=admin -pa=admin halt /dev/null 21 rval=$? ;; *) echo usage: $0 {start|stop|start_msg|stop_msg} rval=1 ;; esac exit $rval Scott, Brian [EMAIL PROTECTED] To:[EMAIL PROTECTED] OM cc: Sent by: Subject: ADSM: Dist Re: TSM 5.1 HP-UX 11.11 Startup/Shutdown Scripts Stor Manager [EMAIL PROTECTED] RIST.EDU 08/14/02 05:02 PM Please respond to ADSM: Dist Stor Manager Bob, FYI... Found out this is a bug in the system. APAR is IC34371 and should be fixed in 5.1.5.0. Root cause is still unknown. Regards, Brian Scott EDS - BUR Engineering Enterprise Distributed Capabilities MS 3278 Troy, MI 48098 * phone: 248-265-4596 (8-365) * mailto:[EMAIL PROTECTED] -Original Message- From: Bob Booth - UIUC [mailto:[EMAIL PROTECTED]] Sent: Tuesday, August 13, 2002 3:36 PM To: [EMAIL PROTECTED] Subject: Re: TSM 5.1 HP-UX 11.11 Startup/Shutdown Scripts What signal is killproc() sending to the server main thread? The server will shutdown gracefully if it gets a SIG 15. See if you send a SIG 15 and the lock file goes away. Just a try. bob On Tue, Aug 13, 2002 at 03:31:26PM -0400, Scott, Brian wrote: Hey gang, Does anyone have a sample startup/shutdown script for the TSM server on HP? On the shutdown the bourne shell will run a killproc on dsmserv process but the adsmserv.lock file under /opt/tivoli/tsm/server/bin doesn't get deleted. HP doesn't reuse the lock file when you try to restart TSM so I have to delete it every time. Anyone come across this on TSM 5.1? Thanks, Brian Brian Scott EDS - BUR Engineering Enterprise Distributed Capabilities MS 3278 Troy, MI 48098 * phone: 248-265-4596 (8-365) * mailto:[EMAIL PROTECTED]
Re: Another TSM mystery...
Is passwordaccess generate in the dsm.sys (or dsm.opt if NT) for this client? If so, it may be a known problem that's been discussed here before... the client and server will generate a new password periodically, and it is kept encrypted in a file on the client. Occasionally, the generated password will start with a hex '00', which causes that behavior. I'm not sure if it must start with '00', or only contain it, and I'm not sure if it's the encrypted or un-encrypted form that contains the '00'... but we have had this happen, and had to reset the password at the server, then manually run a client session and answer the password prompt with the new password... HTH Robin Sharpe Berlex Labs Hunley, Ike Ike.Hunley@B CBSFL.COMTo:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 06/17/02 Subject: 05:05 PM Another TSM mystery... Please respond to ADSM: Dist Stor Manager The TSM server received this message 2 weeks ago, last week and today ANR0424W Session 39912 for node HOC0843P12 (TDP Infmx AIX42) refused - invalid ANR0424W password submitted. Each time we reset the password, issuing UPD NODE nodename pswd PASSEXP=0 which should make this a non- expiring password. How does a perfectly good password, suddenly invalid(no one made changes they cared to admit to...)? What would make a non-expired password expire? Blue Cross Blue Shield of Florida, Inc., and its subsidiary and affiliate companies are not responsible for errors or omissions in this e-mail message. Any personal comments made in this e-mail do not reflect the views of Blue Cross Blue Shield of Florida, Inc.
Re: Licencing. (and pricing)(and support)
So then, WHY do some Tivoli developers have @ibm.com in their email addresses? Robin Sharpe Berlex Labs Mark Stapleton stapleto@BER To:[EMAIL PROTECTED] BEE.COM cc:(bcc: Robin Sharpe/WA/USR/SHG) Subject: 03/06/02 Re: Licencing. (and pricing)(and support) 11:06 AM Please respond to ADSM: Dist Stor Manager Late response, I'm afraid, but nonetheless true. On Wed, 16 Jan 2002 10:28:16 -0500, it was written: And WHY do I have to call a DIFFERENT sales person for Tivoli software than IBM software? It's a different company. And WHY do I have to have a DIFFERENT support contract for Tivoli software than IBM software? It's a different company. And WHY do I have to call a DIFFERENT support number for Tivoli problems than IBM software? It's a different company. And WHY do I have to have a DIFFERENT logon for the Tivoli data base than for IBMLINK, when the APARs show up in IBMLINK eventually anyway? It's a different company. ;o) -- Mark Stapleton ([EMAIL PROTECTED])
Re: Server Migration - AIX to HP
Bob, Don't bother. I can tell you from experience that it won't work, and Tivoli support will tell you it isn't supported. BTW, export/import isn't supported either, although it may work. We are in the middle of a migration from AIX to HP (L2000 server). Tried the same thing. The problem is apparently in the way TSM records data on tape on AIX vs. on HP-UX TSM on HP-UX could not read the tape created on AIX. I was able to export the database and import on HP-UX. But, like you, we have a lot of node data on AIX (about 60TB). By my rough calculations, (using DLT7000 drives) it would have taken more than six months of exporting, 24x7, optimistically assuming a 5MB/sec throughput. What we are now doing is keeping our old TSM and library online until the old data expires which will be about a year. Then we'll export/import any remaining archives. Luckily for us, we have the hardware to do this. Some other folks have suggested FTP'ing the database from AIX to HP-UXor 'tar'ing it... I didn't think of it, and now it's too late to try, since we've been backing up on HP for several months. I also didn't try the unloaddb/loaddb, but I suspect the same problem would exist. Good Luck Robin Sharpe Berlex Labs ARhoads arhoads@PACB ELL.NET To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 03/01/02 Subject: 08:32 AM Re: Server Migration - AIX to HP Please respond to ADSM: Dist Stor Manager Bob, the issue of 'it won't work' doesn't preclude you from trying it! It should work since the platforms are both UNIX (handling tape read/write sim.). Otherwise you'd unloaddb to a file and then move the file to the new platform -- maybe preferable anyway to re-optimize the database... Steffan - Original Message - From: Smith, Bob [EMAIL PROTECTED] To: [EMAIL PROTECTED] Sent: Friday, March 01, 2002 12:51 AM Subject: Server Migration - AIX to HP Has anyone successfully moved a TSM server from one platform to another, copying the database via BACKUP DB and RESTORE DB commands? We are looking at moving from AIX (RS-6000) to HP (N-Class) in this way, but IBM are suggesting that this won't work as the TSM database might be platform-dependent. They suggest using export/import for the all the nodes, but we have 23 Tb of backup data so that would be impossible. IBM 3494 with 3590E drives is used to hold the data. Thanks Bob Smith - Core Infrastructure EMEA EDS UK c/o Rolls-Royce plc, Derby UK tel: 01332 522029 email: [EMAIL PROTECTED] - EDS userid LZTZ3V
Re: Server Migration - AIX to HP
Yes, that was from Tivoli support, although if you push hard enough you could probably get some help. But this is what they told me: You can try exporting the server from AIX and importing it on HP-UX. It might work, but is not supported. Maybe they meant a full import/export -- server and all nodes. And they could have been misinterpreting something. Here's another example of support misleading us: out NT group is testing W2K and had some trouble rebuilding a server from bare metal... support told them TSM does not support bare metal restore on Windows. What they should have said (I think) is TSM does have an automatic bare metal restore for Windows. The NT guys started panicing... I directed them to the books and help screens on the client, and they were able to figure out how to do it. Robin Sharpe Berlex Labs Richard Cowen richard_cowe [EMAIL PROTECTED]To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 03/01/02 Subject: 02:17 PM Re: Server Migration - AIX to HP Please respond to ADSM: Dist Stor Manager -Original Message- From: Robin Sharpe [mailto:[EMAIL PROTECTED]] Sent: Friday, March 01, 2002 1:40 PM To: [EMAIL PROTECTED] Subject: Re: Server Migration - AIX to HP ... BTW, export/import isn't supported either, although it may work. ... Is this from a Support call? The MVS,NT,AIX, and HPUX TSM Guides at least imply cross-platform import is supported: Importing Data from Sequential Media Volumes Before you import data to a new target server, you must: 1. Install TSM on the target server. This step includes defining disk space for the databas and recovery log. For information on installing TSM, see Quick Start. 2. Define server storage for the target server. Because each server operating system handles devices differently, TSM does not export server storage definitions. Therefore, you must define initial server storage for the target server. TSM must at least be able to use a drive that is compatible with the export media. This task can include defining libraries, drives, device classes, storage pools, and volumes. See the Administrator's Guide that applies to the target server.
Re: Server Migration - AIX to HP
In our case, it was a combination of company policy and $$$. We have a global contract with HP through our parent. We were looking to expand our library (ATL P3000) because we had overflowed it. I looked into upgrading it to a P6000 (basically another P3000 bolted on with a pass-through port), but the cost was pretty high. next step was to look at the HP 20/700 (really a STK L700)... which offers more slots than the P6000 upgrade for quite a bit less cost. Problem was, HP would not support their lib on an IBM box. So we went around and around with HP Tivoli on the support issue, and finally decided to get the 20/700 and a new HP L2000 server... then we planned to move the RS6000/P3000 to another location and set up server to server. The migration problem has delayed those plans, but we should get there eventually. So far, The 20/700 has been OK, but there do seem to be some strange I/O errors at times, and the lib's operator interface isn't as advanced as the P3000. The P3000's web interface only works on Solaris and NT, so we can't use it. The HP web interface is built into the lib, but it doesn't allow you to move carts around. Oh well. Robin Sharpe Berlex Labs Thomas A. La Porte [EMAIL PROTECTED] WORKS.COM To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 03/01/02 02:26 PMSubject: Please respond to Re: Server Migration - AIX to HP ADSM: Dist Stor Manager Just out of curiosity for the several people who have mentioned moving from AIX to HP/UX, what has been the driving motivation to switch platforms? We've been completely satisfied with AD/TSM on AIX for the past six years, in spite of having little to no AIX expertise in house (our TSM servers are the only AIX boxes in our environment). Just wondering what prompts the shift? -- Tom Thomas A. La Porte DreamWorks SKG [EMAIL PROTECTED] On Fri, 1 Mar 2002, Robin Sharpe wrote: Bob, Don't bother. I can tell you from experience that it won't work, and Tivoli support will tell you it isn't supported. BTW, export/import isn't supported either, although it may work. We are in the middle of a migration from AIX to HP (L2000 server). Tried the same thing. The problem is apparently in the way TSM records data on tape on AIX vs. on HP-UX TSM on HP-UX could not read the tape created on AIX. I was able to export the database and import on HP-UX. But, like you, we have a lot of node data on AIX (about 60TB). By my rough calculations, (using DLT7000 drives) it would have taken more than six months of exporting, 24x7, optimistically assuming a 5MB/sec throughput. What we are now doing is keeping our old TSM and library online until the old data expires which will be about a year. Then we'll export/import any remaining archives. Luckily for us, we have the hardware to do this. Some other folks have suggested FTP'ing the database from AIX to HP-UXor 'tar'ing it... I didn't think of it, and now it's too late to try, since we've been backing up on HP for several months. I also didn't try the unloaddb/loaddb, but I suspect the same problem would exist. Good Luck Robin Sharpe Berlex Labs ARhoads arhoads@PACB ELL.NET To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 03/01/02 Subject: 08:32 AM Re: Server Migration - AIX to HP Please respond to ADSM: Dist Stor Manager Bob, the issue of 'it won't work' doesn't preclude you from trying it! It should work since the platforms are both UNIX (handling tape read/write sim.). Otherwise you'd unloaddb to a file and then move the file to the new platform -- maybe preferable anyway to re-optimize the database... Steffan - Original Message - From: Smith, Bob [EMAIL PROTECTED] To: [EMAIL PROTECTED] Sent: Friday, March 01, 2002 12:51 AM Subject: Server Migration - AIX to HP Has anyone successfully moved a TSM server from one platform to another, copying the database via BACKUP DB and RESTORE DB commands? We are looking at moving from AIX (RS-6000) to HP (N-Class) in this way, but IBM are suggesting that this won't work as the TSM database might be platform-dependent. They suggest using export/import for the all the nodes, but we have 23 Tb of backup data so that would be impossible. IBM 3494 with 3590E drives is used to hold the data. Thanks Bob Smith - Core Infrastructure EMEA EDS UK c/o Rolls-Royce plc, Derby UK tel: 01332 522029 email: [EMAIL PROTECTED] - EDS userid LZTZ3V
Re: Looking for sites with TSM 4.X running on HP-UX server
Mark (and anyone else monitoring), Update on our situation, which I think was similar to yours... We've been having erratic TSM server hangs, but only the TSM server, the OS and other minor apps on the system are OK. There were no evident bottlenecks or problems. Yesterday, I reviewed the actlog for the previous 4 days and saw many I/O errors like this: 02/18/02 12:04:05 ANR8302E I/O error on drive DRIVE4 (/dev/rmt/11m) (OP=FSR, CC=-1, KEY=FF, ASC=FF, ASCQ=FF, SENSE=**NONE**, Description=An undetermined error has occurred). Refer to Appendix D in the 'Messages' manual for recommended action. I searched the actlog for mount msgs, ANR8302E msgs, and dismount msgs. Whenever tape volume 000432 was mounted, it was followed by many 8302 msgs, and never saw a dismount for that tape. Conclusion: there must be something wrong with that tape! I marked it unavailable and have not seen the hang in the last 24 hours. We know that tape was used a few days prior to the hangs for backup of an NT server that had known disk (FAT) corruption, but we thought it was in the recycle bin which is excluded. We were then constantly doing storage pool backups which we think mounted that tape and caused the hang when it reached the corruption on the tape. I am now concerned that an apparent media fault can cause TSM to hang. I have an open PMR with Tivoli. I guess the moral of the story is -- investigate I/O errors thoroughly! Robin Sharpe Berlex Labs
Re: Looking for sites with TSM 4.X running on HP-UX server
Mark, We are running TSM 4.1.4.1 on an L2000-5x, HP-UX 11.00. We have an HP 20/700 library with 10 DLT8000 drives, each on its own SCSI bus. We are in the early stages of a SAN implementation (with HP XP512 disk) and have just added 2 A5158A fiber cards to the L2000. At about the same time we installed them, we noticed erratic hangs on the TSM server, which eventually clear up... but the effect on TSM operations has been pretty bad. Just applied some patches to HP-UX about an hour ago... will let you know if it fixes things. What is your specific problem? What is you environment (server, library, etc.)? Robin Sharpe Berlex Labs Kovacs, Mark Mark.Kovacs@P HARMA.COM To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 02/11/02 05:15 Subject: PMLooking for sites with TSM 4.X running on HP-UX server Please respond to ADSM: Dist Stor Manager To all: I would like to get in contact with anyone running TSM 4.X on HP-UX 11.00. We have multiple servers and one is experiencing application issues. We have been in continuous contact with TSM support, but find it very frustrating. We believe it may be a combination of patches on both TSM as well as HPUX. Any and all help would be greatly appreciated. thanks, mark
Re: Looking for sites with TSM 4.X running on HP-UX server
Ok Mark, Here is my swlist -l product: (See attached file: swlist)(let me know if you can't get the attachment, I'll insert the text). We haven't had such serious problems (no corruption). 4.1.4.1 has been pretty stable for us... until last week. So far, after the patches this morning, it looks pretty good. I had heard that the earlier mods of 4.1 were buggy... maybe that is at the root of your trouble? We have about 26 unix (HP and Sun) and 50 NT clients. We also have another TSM on an IBM F50 with an ATL P3000, which we are migrating from. We plan to move it to another location after migration. Our L2000 has two 550Mhz procs, and 2GB RAM. Database and disk storage pools are on two HP 2100 arrays... we plan to move the database and logs to the XP512 when it is online. Good Luck Robin swlist Description: Binary data
Phantom dsmserv process on HP-UX , TSM 4.1.4.1
Hi all, Has anyone seen something like this? wau026:/home/root ps -ef|grep dsm root 5364 1 0 Feb 1 ?1120:55 ./dsmserv -quiet root 6597 0 0 09:40:55 ? 0:00 ./dsmserv -quiet www 6714 6713 0 09:41:58 ? 0:00 dsmadmc -server=localsrv -id=admin -pa=admin -comma select s www 6713 6512 0 09:41:58 ? 0:00 sh -c dsmadmc -server=localsrv -id=admin -pa=admin -comma s That's TSM 4.1.4.1 on an HP-UX 11.0 system. It happens quite often... we have a phantom dsmserv process that we can't kill. TIA Robin Sharpe Berlex Labs
Re: Background / Foreground Performance Mystery
Curious indeed... did you try falling back to the older client? If you do, i would suggest setting up a temporary node on the server and test under that name... there may be some incompatibilities between backups from different client levels, and you probably don't want to be flipping back and forth... you may not be able to restore data backed up by the new client with the old one. Robin Sharpe Berlex Labs Seay, Paul seay_pd@NAPT HEON.COM To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/30/02 Subject: 08:02 AM Re: Background / Foreground Performance Mystery Please respond to ADSM: Dist Stor Manager What Resourceutilization number are you using? -Original Message- From: Magura, Curtis [mailto:[EMAIL PROTECTED]] Sent: Wednesday, January 30, 2002 5:10 AM To: [EMAIL PROTECTED] Subject: Re: Background / Foreground Performance Mystery Tried Wanda's idea last night. Backup completed in what we expect the normal time to be. This is using my logon (admin auth on the local NT resource domain) and opening a window with the command dsmc sched. Not sure what this tells us though I'm thinking this is still in the foreground vs. an NT service? Starting to get way over my NT skills at this point! I've asked some of the local Intel support folks about dispatching priorities and if/where they can be set. Are any of the NT savvy folks on the list aware of anyplace that can be set on NT4? Curt Magura Lockheed Martin EIS Gaithersburg, Md. 301-240-6305 -Original Message- From: Prather, Wanda [mailto:[EMAIL PROTECTED]] Sent: Tuesday, January 29, 2002 2:08 PM To: [EMAIL PROTECTED] Subject: Re: Background / Foreground Performance Mystery I have no clue; but one difference in running dsmc as yourself and via the scheduler is that you are running under different accounts. What happens if you run the scheduler under your own account? -Original Message- From: Magura, Curtis [mailto:[EMAIL PROTECTED]] Sent: Tuesday, January 29, 2002 11:06 AM To: [EMAIL PROTECTED] Subject: Background / Foreground Performance Mystery Client: O/S - NT4 SP6A TSM 4.2.1.20 Server: AIX 4.3.3 ML6 TSM 4.2.1.7 We have one NT client the has degraded terribly over the past couple of months. We are in the midst of upgrading clients to .20 and the hope was it would help. No luck. Here's the weird part. If I logon to the machine and run dsmc incr the backup performance is about what we expect. Last night we backed up 155.56 GB in 09:05:16. This is an Intel file server with a tad over 4.8 million files on it. If the TSM Scheduler Service starts the backup it takes more than 24 hours to complete! And even more weird it used to run using the scheduler just fine. Standard caveatsnobody has changed anything or so they all say!!! Except of course the upgrade to the newer client that I mentioned. I also deleted and recreated the TSM Scheduler Service No difference. The machine is set give performance preference to background tasks per right clicking on properties of My Computer. I'm at a loss as to why the scheduler performance in unacceptable. Earlier today I setup an AT job to start up the dsmc incr command to see how it would run. Appears to be running at the same pace as if it was started via the Scheduler service. I'll take any help on this one! Curt Magura Lockheed Martin EIS Gaithersburg, Md. 301-240-6305
Re: A shot in the dark? Here goes...
I don't think it's possible; certainly not supported. Guess it wouldn't hurt to try, though. It was a great movie, though.;) Robin Sharpe Berlex Labs TSM Group tsmgroup@HOT MAIL.COM To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/25/02 Subject: 04:05 PM A shot in the dark? Here goes... Please respond to ADSM: Dist Stor Manager We're running an AIX 4.3 server, TSM server version 4.2.1.8 And unfortunately one of our clients is an HP UX 10.20 Box. We have tested and have successfully backed up / restored a file using this combination but it is slow. Has anyone successfully loaded the client for HPUX 11.x on a HPUX 10.20 Box? _ Join the world s largest e-mail service with MSN Hotmail. http://www.hotmail.com
Re: I surrender ? : HOW CAN I GET THE TSM SUPPORT ?
Have you registered for Tivoli's online support? I have used it to enter problems and they have always responded within a reasonable time. You can also correspond with them, and track their progress and the status of the call. Robin Sharpe Berlex Labs Michel David moi_md@YAHOO .COM To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/21/02 Subject: 01:34 PM I surrender ? : HOW CAN I GET THE TSM SUPPORT ? Please respond to ADSM: Dist Stor Manager I think than the guys there are not better than Richard, Wanda, PAul Demetrius and other friends that alwais solve every problem. We ARE ready to PAY. It's not the problem. The problem is that we didn't find a way to get this support. I mean that I want a man and not a stupid automatic mail responser. Someone has an idea ? Thank you. Michel MARNET --- Malbrough, Demetrius [EMAIL PROTECTED] wrote: Are your DB Recovery Logs mirrored? If so, you can try to bring the server up by reading the mirrors. In your dsmserv.opt file change MIRRORREAD LOG to Verify! Otherwise, call TSM Support!!! -Original Message- From: Lloyd Dieter [mailto:[EMAIL PROTECTED]] Sent: Monday, January 21, 2002 10:22 AM To: [EMAIL PROTECTED] Subject: Re: Please HELP! Restore DB without DBBACKUP tape. Michael, To the best of my knowledge, you MUST have a valid DB backup tape to recover. AFAIK, it is not possible to read storagepool media without an intact database. If your DB is not intact, and your DB backup tapes are no good, I would recommend that you call TSM support to see if they can help you. Good luck! -Lloyd On Mon, 21 Jan 2002 07:19:35 -0800 Michel David [EMAIL PROTECTED] wrote: Hi everybody! Thank you Paul. I try once more to get help to our powerfull List. This is the biggest disaster I can imagine. That looks like No DBBACKUP tape are available ! Is it possible to recover the data FROM the POOL TAPES, if I know which tape is right pool tape ? This is a real emmergency. We tried everything else I think. Thank you. Michel DAVID MARNET __ Do You Yahoo!? Send FREE video emails in Yahoo! Mail! http://promo.yahoo.com/videomail/ -- - Lloyd Dieter- Senior Technology Consultant Synergy, Inc. http://www.synergyinc.cc [EMAIL PROTECTED] Main:716-389-1260fax:716-389-1267 - __ Do You Yahoo!? Send FREE video emails in Yahoo! Mail! http://promo.yahoo.com/videomail/
Re: One Client, Two File Types, Different Retention Reqd.
I don't see any way around using include/exclude lists. What is the objection to using them? In your case, you wouldn't be excluding anything, but including specific files to a different MC. Also, there are some files you definitely do not want to be backing up... like dsmerror.log Robin Sharpe Berlex Labs Paul Biggin paul.biggin@CONS IGNIA.COMTo:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/10/02 12:50 PM Subject: Please respond toOne Client, Two File Types, Different Retention Reqd. ADSM: Dist Stor Manager Hi, Can someone please suggest the best/preferred way of having different retention periods for different groups of files from one client. I've got a client that has two different groups of files, lets say /data and /test. /data needs to be retained for one month, /test for one week. I assume that to have each group held for different retention periods, I will have to set up two management classes each pointing to their own backup copy groups. But how do I associate the files with the intended management class/copy group? The client is defined to a single policy domain. I can't try include/exclude lists, as we don't use them! Is there another way? Cheers Paul This email and any attachments are confidential and intended for the addressee only. If you are not the named recipient, you must not use, disclose, reproduce, copy or distribute the contents of this communication. If you have received this in error, please contact the sender and then delete this email from your system.
Re: Handling spikes in storage transfer
Do you still have the older sessions in your actlog? Or you could look in the summary table. Try to see if it's because of a lot more files, or just more data. If not many more files, look in the contents table for files larger than 30GB. If none found, gradually decrease the size until you find some. That's what I'd do. Those queries might take some time, though. Robin Sharpe Berlex Labs Zoltan Forray/AC/VCU To:[EMAIL PROTECTED] zforray@VCU. cc:(bcc: Robin Sharpe/WA/USR/SHG) EDU Subject: Handling spikes in storage transfer 01/14/02 09:47 AM Please respond to ADSM: Dist Stor Manager I have an SGI client node that while normally sends = 1.5GB of data, is now sending 36GB+. Without accessing the client itself, how can I find out what is causing this increase in TSM traffic ? I have contacted the client owner, but their response is taking too long and this spike it wreaking havoc on TSM. Zoltan Forray Virginia Commonwealth University - University Computing Center e-mail: [EMAIL PROTECTED] - voice: 804-828-4807
Re: Copygroups
That's Ok as long as you never do more than one backup per day. If for example, you do a second backup, VEREXISTS=60 will cause the oldest one from 60 days ago to be deleted. To be absolutely safe, you need to code VEREXISTS= and VERDELETED=NOLIMIT. This is documented in the Administrator's Guide in a section dealing with PIT restores. Robin Sharpe Berlex Labs Vint Maggs vint.maggs@S RS.GOV To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/14/02 Subject: 12:59 PM Copygroups Please respond to ADSM: Dist Stor Manager How do you recommend setting up a management class that would allow me to retain backups for 60 days while allowing a PIT restore to any date within the past 60 days? I am thinking I should set the copygroup parameters VERExists, VERDeleted, RETExtra and RETOnly to 60. Does this seem correct? Thanks, Vint
Re: Handling spikes in storage transfer
Z. Is that list latest first or earliest first? Either way, you went from about 1000 files one day to 36,806 the next day... But still, there are four sessions with about 1000 files and nearly 40 GB. So, I'd try looking at the contents table. I think that's the only table on the server that captures individual files' sizes. If I had to guess what was going on, I'd say maybe the client had a lot of activity one day that created a very large log file, which probably never gets pruned. You might ask around if anything strange happened on the client in the 24 hours before the first big backup Robin Sharpe Berlex Labs Zoltan Forray/AC/VCU To:[EMAIL PROTECTED] zforray@VCU. cc:(bcc: Robin Sharpe/WA/USR/SHG) EDU Subject: Re: Handling spikes in storage transfer 01/14/02 03:03 PM Please respond to ADSM: Dist Stor Manager I am pretty sure it isn't a growth in the # of files. While the # of files backed-up did spike a little, it doesn't correspond with the bytes transfered. I think it is one or two big files. Here is a snapshot from the SMF ACCOUNTING RECORDS generated by the server (OS390) for the past 8-days. Objects Total Data Total KB Inserted Sent/Recvd Backup/Arch - 16,097 1,136,204 1,131,671 648264,535 264,354 961368,767 368,497 36,806 40,531,165 40,516,245 1,034 39,270,046 39,265,010 1,096 39,469,387 39,464,295 1,309 39,418,974 39,412,931 2,033 39,563,065 39,557,509 --- Zoltan Forray Virginia Commonwealth University - University Computing Center e-mail: [EMAIL PROTECTED] - voice: 804-828-4807 Robin Sharpe [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 01/14/2002 12:11 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Subject:Re: Handling spikes in storage transfer Do you still have the older sessions in your actlog? Or you could look in the summary table. Try to see if it's because of a lot more files, or just more data. If not many more files, look in the contents table for files larger than 30GB. If none found, gradually decrease the size until you find some. That's what I'd do. Those queries might take some time, though. Robin Sharpe Berlex Labs Zoltan Forray/AC/VCU To:[EMAIL PROTECTED] zforray@VCU. cc:(bcc: Robin Sharpe/WA/USR/SHG) EDU Subject: Handling spikes in storage transfer 01/14/02 09:47 AM Please respond to ADSM: Dist Stor Manager I have an SGI client node that while normally sends = 1.5GB of data, is now sending 36GB+. Without accessing the client itself, how can I find out what is causing this increase in TSM traffic ? I have contacted the client owner, but their response is taking too long and this spike it wreaking havoc on TSM. Zoltan Forray Virginia Commonwealth University - University Computing Center e-mail: [EMAIL PROTECTED] - voice: 804-828-4807
Re: Restoring multiple filesets from command line.
Simon, I was about to say no way, but I poked around a little, and there is indeed a way to do it. (I really didn't think there was!). You have to use the -filelist option on the restore comand. For example, I did this: wau026:/home/root dsmc res -filelist=/home/root/reslist Tivoli Storage Manager Command Line Backup Client Interface - Version 4, Release 2, Level 1.0 (C) Copyright IBM Corporation, 1990, 2001, All Rights Reserved. Restore function invoked. Node Name: WAU026 Session established with server TSM_WAYNE: HP-UX Server Version 4, Release 1, Level 4.1 Server date/time: 01/10/02 11:02:14 Last access: 01/09/02 22:03:16 ** Interrupted ** ANS1114I Waiting for mount of offline media. Restoring 2,890 /apps/adsm.sh [Done] Restoring 3,687 /home/root/autosum [Done] Total number of objects restored: 2 Total number of objects failed: 0 Total number of bytes transferred: 6.47 KB Data transfer time:0.00 sec Network data transfer rate:58,868.96 KB/sec Aggregate data transfer rate: 0.03 KB/sec Elapsed processing time: 00:03:28 where /home/root/reslist is a file I created and contains this: /home/root/autosum /apps/adsm.sh On my system, /home/root and /apps are two different unix filesystems. As you can see, it worked. The bad news is, you can't put wildcards inside the filelist... all files must be fully qualified. And TSM restores them in the order specified, so some tape juggling may result. Now, if we could only find a way to get TSM to give us a list of files sorted in the order they are stored on tape, we could implement that optimized node restore command that I wished for! Robin Sharpe Berlex Labs Pope, Simon R. spope@WESTERNPO WER.CO.UK To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/10/02 07:08 Subject: AM Restoring multiple filesets from command line. Please respond to ADSM: Dist Stor Manager Hi All, Does anyone know if it is possible to specify multiple filesystems for restore using a single command line instruction. It can be done from the GUI very easily but we can't figure how it can be done from the command line (or even if it can be done!). TSM Server 4.1.3 Clients 4.1.2 all running on AIX 4.3.3 Thanks Simon Pope _ This email and any files transmitted with it are confidential and intended solely for the use of the individual or entity to whom they are addressed. If you have received this email in error please notify [EMAIL PROTECTED] This footnote also confirms that this message has been swept for all known viruses by the MessageLabs Virus Scanning Service. For further information please visit http://www.messagelabs.com
Re: library sharing without SAN
Does a configuration such as that need special support by the library? What happens if both systems send commands to the robot? Also, what about the drives? Most SCSI libraries I know of have the drives cabled directly on their own SCSI channels. Definitely the two I have... an ATL P3000 and an HP 20/700. This sounds very interesting, and I'm wondering if I can use it while we migrate from the P3000 to the 20/700. Currently they are attached to two different servers - the P3000 to an F50 and the 20/700 to an HP L2000. As far as I know, the method of recording data on DLT tapes differs on those two OSes, and they cannot read each others' tapes. Robin Sharpe Berlex Labs Bill Smoldt smoldt@STORS OL.COM To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/09/02 Subject: 08:54 AM Re: library sharing without SAN Please respond to ADSM: Dist Stor Manager You can connect a library on a SCSI bus between the two systems and share the library the same way as you would in a SAN. I've never tried it between Solaris and Windows, but it should work fine as long as the controllers don't interfere with each other electrically (they will typically be different brands). The controllers must have different SCSI target addresses - I usually use the default of 7 on one and 6 on the other. If you just want to get some experience with shared libraries within TSM, the simplest method is to create a second TSM server instance on the same system. Then you can define a shared library between the two. I use this method for training classes. Bill Smoldt STORServer, Inc. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED]]On Behalf Of Chandrasekhar CR Sent: Wednesday, January 09, 2002 1:54 AM To: [EMAIL PROTECTED] Subject: library sharing without SAN hello , For experimental purpose can we use autochanger for sharing between two TSM servers ( solaris and windows ) without SAN ,if yes then what is the procedure to do it. if any one know about this please help me. TSM administrator
Re: library sharing without SAN
Now that I think about it, I'm not sure what advantage we would gain, if any. Our goal is to eventually move the P3000 and f50 to another location anyway. Wait a minute, I just thought of a benefit... the reason we got the new library was because we overflowed the old one. It's still a problem juggling tapes just to do reclaims on the old system. Anyway, I was just curious how it would all work... I didn't think you could define SCSI libraries to TSM as SHARED (or don't you have to?). How do you physically cable it? Just replace the SCSI terminator(s) with a cable(s) to the other system? I guess I would want to something like this (please excuse rudimentory non-graphics): HP L2000 (TSM2) -scsi- HP 20/700 -scsi IBM F50 (TSM1) -scsi- ATL P3000 Then I would want to check some tapes into the 20/700 to allow TSM1 to do reclaims. 20/700 has 553 available slots; P3000 has none. Robin Sharpe Berlex Labs Bill Smoldt smoldt@STORS OL.COM To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/09/02 Subject: 10:15 AM Re: library sharing without SAN Please respond to ADSM: Dist Stor Manager No special support with all the SCSI adapters that I've dealt with - mostly Adaptec and QLogic. The hardware has been capable of this for years and I've used this on many OpenVMS clusters and NT clusters. The initiators don't talk to each other, and the devices have to be capable of doing a SCSI disconnect. All arbitration and locking for the robot and drives takes place in the TSM Server. Might ease your library movement, but I've never tried crossing the data tapes between platforms. It's certainly unsupported so you'd want to rewrite the tapes even if it worked. Is that what you're thinking? Bill Smoldt STORServer, Inc.
Re: Tape Drive Failures
Looks to me like you need to do an AUDIT LIBRARY libname CHECKL=BARCODE. TSM keeps its own library inventory. If someone opens the library and removes a tape or moves them around, TSM's inventory will be out of sync with what's really in there. This can also happen on some libraries if the library auto-loads tapes through the entry-exit ports while tapes are in drives... the new tapes may go into the home slot of a mounted tape... happened to us on our ATL P3000, so we turned off auto-load. Robin Sharpe Berlex Labs Bill Wheeler Bill.Wheeler@LA- Z-BOY.COMTo:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 01/09/02 03:03 PM Subject: Please respond toRe: Tape Drive Failures ADSM: Dist Stor Manager Here is a look at what the message is giving me in the actlog when it fails. It looks just like a plan tape error: ANR8499I Command accepted. 01/05/02 08:58:03 ANR8300E I/O error on library MAGSTAR (OP=6C03, CC=314, KEY=05, ASC=3B, ASCQ=0E, SENSE=70.00.05.00.00.0- 0.00.48.00.00.00.00.3B.0E.FF.02.00.A8.00.00.1E.09.01.00- .00.00.00.00.00.00.A5.00.00.00.00.00.00.00.00.00.00.00.- 00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.00.0- 0.00.00.00.00.00.00.00.00.00.00.00.32.38.37.20.20.20.20- .00., Description=The source slot or drive was empty in an attempt to move a volume). Refer to Appendix B in the 'Messages' manual for recommended action. 01/05/02 08:58:03 ANRD mmsscsi.c(10695): Could not move volume 1DA7D from slot 31 to slot 16. 01/05/02 08:58:03 ANR8841I Remove volume from slot 31 of library MAGSTAR at your convenience. After this the drive then goes offline. So when the archive runs at 22:30 this is the message it gets: 01/05/02 22:41:20 ANR8447E No drives are currently available in library MAGSTAR. 01/05/02 22:41:21 ANR1404W Scratch volume mount request denied - mount failed. When i look at the error message in the O/S errpt this is what i see 4865FA9B 0108012102 P H rmt1 TAPE OPERATION ERROR 476B351D 0108012102 P H rmt1 TAPE DRIVE FAILURE 0F78A011 0108011902 T H rmt1 RECOVERY LOGIC INITIATED BY DEVICE Any ideas??? Bill Wheeler PDM Administrator La-Z-Boy Incorporated (734) 242-1444 x 6170 -Original Message- From: George Lesho [mailto:[EMAIL PROTECTED]] Sent: Wednesday, January 09, 2002 1:41 PM To: [EMAIL PROTECTED] Subject: Re: Tape Drive Failures Perhaps if y ou included some representitive samples of the errors so the SCSI error codes there is some hope of helping you with the problem. Pull some from your O/S errpt and TSM actl. Cheers George Lesho AFC Enterprises Storage/System Admin Bill Wheeler [EMAIL PROTECTED]@VM.MARIST.EDU on 01/09/2002 10:20:52 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] To: [EMAIL PROTECTED] cc:(bcc: George Lesho/Partners/AFC) Fax to: Subject: Tape Drive Failures Hello *SMers, Just a quick question out there for all of you. I am currently receiving tape drive failures when trying to check in and out tapes, also during our nightly archive. They have been increasing during the last few months. My environment is Magstar 3570, 2 drives. I am currently running TSM 4.1.2.0 on AIX 4.3.3. I have been cleaning the drives regularly, but am not sure if I should be IBM involved or not. One of the main problems is that when the drive gets the error, the drive does not reset. It could be the tapes causing the problems, but the drives are still not resetting. Just yesterday someone was checking in tapes and loaded the wrong one, but the drive failed due wrong tape being mounted. Then the drive did not reset itself. Is there anything that I can do without calling IBM to see if I can get this fixed? Thanks in advance, Bill Wheeler PDM Administrator La-Z-Boy Incorporated (734) 242-1444 x 6170
Re: Please Explain (again)
Well, how to lookup individual sessions: The short answer is SELECT * FROM SUMMARY WHERE ENTITY='nodename' AND ACTIVITY='BACKUP' Multi streamed backups are a great enhancement, but there are some subtleties. With a single streamed backup, you have one session and one statistics report on standard output (if running from a script). If running from a unixcommand line it comes out on your terminal screen. If running from a GUI it is in a window, but has a slightly different format (if I remember correctly). With a multi streamed backup, you have several sessions, as you have stated. But you still get a single statistics report. In the GUI you can see the individual sessions, but on the command line version, there is no indication that it's a multi stream. And this where I think the problem lies. So, when you do the select from summary, you should get several records that start at the same time (or thereabouts). These would be the sessions that make up the multi streamed backup. I think you have to calculate the elapsed time (end-start)... For the ***FLASH*** light bulb just went on over my head! This is why it isn't accurate: Each session on a multi stream backup will be a different elapsed time (probably). Since they are running concurrently, the elapsed time for the backup is of course the end time of the latest session minus the start time of the earliest. But each session has a data transfer time that may mor may not overlap the other sessions. The total data transfer is the sum of those of each session... so it could conceivably be more than the elapsed time. Still doesn't seem to make sense though (light bulb is dimming now and less than an hour of 2001 remains!) I'm really getting curious about this now... guess I'll have to do some tests when I get back to work on Wednesday. Hope everyone has a Safe Happy New Year! Robin Sharpe Berlex Labs Miles Purdy PURDYM@FIPD. GC.CATo:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 12/30/01 Subject: 10:25 AM Re: Please Explain (again) Please respond to ADSM: Dist Stor Manager Hi guys, I do not think any times are wrong nor is there a bug, see example. PS: How do I look up individual sessions for a given backup? Example: Here's how I thought the numbers were arrived at: say processing starts at time zero and runs for 1 hour, 3600 s, just to make it easy, and we backup 100 GB. We also use two streams 50 GB total each, and they run concurrently. %75 of the time is spent sending data, %25 processing overhead. So: total time: 3600 s total bytes: 100 GB aggregate is: 100 * 1024 mb / 3600 = 28 MB /s data transfer time: .75 * 3600 * 2 = 5400 s network throughput is: 100 * 1024 / 5400 = 18 MB / s I think what I meant to say was the network pipe is really _wide_. The network can sustain multiple streams running at full speed. The limiting speed factor may be the server or it may be the network, but I don't think it matters where the speed bottleneck is. So I still don't think it is a bug. Miles -- Miles Purdy System Manager Farm Income Programs Directorate Winnipeg, MB, CA [EMAIL PROTECTED] ph: (204) 984-1602 fax: (204) 983-7557 If you hold a UNIX shell up to your ear, can you hear the C? - [EMAIL PROTECTED] 28-Dec-01 9:56:57 PM As I said a while ago, I think it's a bug. I'm guessing that this is a multi stream backup (I think that has already been established), and the data transfer time is the total of all of the sessions, but the elapsed time is for only one session... can you confirm that Miles? Hopefully you still have records for those sessions in the summary table. Robin Sharpe Berlex Labs Zlatko Krastev/ACIT acit@ATTGLOB To:[EMAIL PROTECTED] AL.NET cc:(bcc: Robin Sharpe/WA/USR/SHG) Subject: 12/27/01 Re: Please Explain (again) 07:30 PM Please respond to ADSM: Dist Stor Manager Sorry, I am taking my words back. Have a look again at the times reported Data transfer time: 8 261.61 sec Elapsed processing time: 01:25:00 This 8261 seconds is definitely much more than 1h25m (5100s) and one of the times is erroneous. Divided by wrong value you're getting one of the rates wrong. Zlatko Krastev IT Consultant Zlatko [EMAIL PROTECTED] 22.12.2001 23:55 Sent by:Zlatko Krastev/ACIT[EMAIL PROTECTED
Re: ATL P2000 dlt 7000 question
Hey Paul, Is this happening on many carts or just a few (or just one)? Have you looked at it... maybe it's broke! Robin Sharpe Berlex Labs Coviello, Paul PCoviello@CM To:[EMAIL PROTECTED] C-NH.ORG cc:(bcc: Robin Sharpe/WA/USR/SHG) Subject: 12/28/01 Re: ATL P2000 dlt 7000 question 01:03 PM Please respond to ADSM: Dist Stor Manager ok I guess I spaced it sorry, a lot is going on. anyways it is very intermittent. I say it is impossible the ATL tech also has said it shouldn't be happening but the operators swear that it is. of course I have no idea how they are put in, or wether there are gremlins involved :-) thanks Paul -Original Message- From: Zlatko Krastev/ACIT [SMTP:[EMAIL PROTECTED]] Sent: Thursday, December 27, 2001 7:31 PM To: [EMAIL PROTECTED] Subject: Re: ATL P2000 dlt 7000 question I also remeber such problem so have searched through my mail list archive folder and really found a thread P2000 questions on 14.09 started by ... Coviello, Paul [EMAIL PROTECTED] and answered by ... Robin Sharpe [EMAIL PROTECTED] :-))) And guess what - the subject was slightly different but the problem was the same, the hint was same :-) Paul, since September did the problem disappeared by itself, is it rarely occuring or was resolved but happened again with no same resolution? Zlatko Krastev IT Consultant Robin Sharpe [EMAIL PROTECTED] on 27.12.2001 22:19:27 Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] To: [EMAIL PROTECTED] cc: Subject:Re: ATL P2000 dlt 7000 question Paul, I remember this same topic being discussed several months ago... I thought it was impossible... how can the library move the slider on the cartridge!? We have a P3000 and have never seen this happen. I don't remember the resolution from that last discussion (or if there was one)... but I'll say again what I said then... call ATL tech support! They are usually pretty helpful. Robin Sharpe Berlex Labs Coviello, Paul PCoviello@CM To:[EMAIL PROTECTED] C-NH.ORG cc:(bcc: Robin Sharpe/WA/USR/SHG) Subject: 12/24/01 ATL P2000 dlt 7000 question 12:40 PM Please respond to ADSM: Dist Stor Manager Hi we have an ATL P2000 with 4 dlt7000's 100 tape capacity. the operations group is seeing DLT tapes coming out of the library write protected. Would anyone care to guess at this? because I can't even come up with a rational explanation nor logical one. :-) I have asked if it is the same tapes, ( going thru list this week) thinking they might have a bad switch but then what moves it! thanks Paul Paul J Coviello Sr Systems Analyst Catholic Medical Center 2456 Brown Ave Manchester NH 03103 (603) 663-5326 [EMAIL PROTECTED]
Re: Please Explain (again)
As I said a while ago, I think it's a bug. I'm guessing that this is a multi stream backup (I think that has already been established), and the data transfer time is the total of all of the sessions, but the elapsed time is for only one session... can you confirm that Miles? Hopefully you still have records for those sessions in the summary table. Robin Sharpe Berlex Labs Zlatko Krastev/ACIT acit@ATTGLOB To:[EMAIL PROTECTED] AL.NET cc:(bcc: Robin Sharpe/WA/USR/SHG) Subject: 12/27/01 Re: Please Explain (again) 07:30 PM Please respond to ADSM: Dist Stor Manager Sorry, I am taking my words back. Have a look again at the times reported Data transfer time: 8 261.61 sec Elapsed processing time: 01:25:00 This 8261 seconds is definitely much more than 1h25m (5100s) and one of the times is erroneous. Divided by wrong value you're getting one of the rates wrong. Zlatko Krastev IT Consultant Zlatko [EMAIL PROTECTED] 22.12.2001 23:55 Sent by:Zlatko Krastev/ACIT[EMAIL PROTECTED] To: ADSM: Dist Stor Manager [EMAIL PROTECTED] cc: Subject:Re: Please Explain If your network capacity greatly exceeds the clients ability to send data then your network rate should greatly exceed the data read (on the node) and write (on the server) rate. Consequently aggregate rate has also to be greatly exceeded. The only explanation I can find to the numbers you're seeing is that the speed of the SP Switch is incorrectly counted in MHz not in MB/s. And later those MHz are converted as usual serial Ethernet in MB/s. Because SP Switch running at low frequency gives high throughput your network rate is displayed wrong. Zlatko Krastev IT Consultant Miles Purdy [EMAIL PROTECTED] on 20.12.2001 17:26:13 Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] To: [EMAIL PROTECTED] cc: Subject:Re: Please Explain I don't think it is a bug. I think is because my network (SP Switch) capacity greatly exceeds the clients ability to send data (even though is an S80). If in effect I'm running 2,3,4,5 backups concurrently. 12/19/01 14:25:28 ANE4952I (Session: 14442, Node: UNXP) Total number of objects inspected: 135 487 12/19/01 14:25:28 ANE4954I (Session: 14442, Node: UNXP) Total number of objects backed up:2 309 12/19/01 14:25:28 ANE4958I (Session: 14442, Node: UNXP) Total number ofobjects updated: 0 12/19/01 14:25:28 ANE4960I (Session: 14442, Node: UNXP) Total number of objects rebound: 0 12/19/01 14:25:28 ANE4957I (Session: 14442, Node: UNXP) Total number of objects deleted: 0 12/19/01 14:25:28 ANE4970I (Session: 14442, Node: UNXP) Total number of objects expired: 13 138 12/19/01 14:25:28 ANE4959I (Session: 14442, Node: UNXP) Total number of objects failed: 0 12/19/01 14:25:28 ANE4961I (Session: 14442, Node: UNXP) Total number of bytes transferred:60.53 GB 12/19/01 14:25:28 ANE4963I (Session: 14442, Node: UNXP) Data transfer time:8 261.61 sec 12/19/01 14:25:28 ANE4966I (Session: 14442, Node: UNXP) Network datatransfer rate:7 682.73 KB/sec 12/19/01 14:25:28 ANE4967I (Session: 14442, Node: UNXP) Aggregate data transfer rate: 12 445.33 KB/sec 12/19/01 14:25:28 ANE4968I (Session: 14442, Node: UNXP) Objects compressed by:0% 12/19/01 14:25:28 ANE4964I (Session: 14442, Node: UNXP) Elapsed processingtime:01:25:00 Miles -- Miles Purdy System Manager Farm Income Programs Directorate Winnipeg, MB, CA [EMAIL PROTECTED] ph: (204) 984-1602 fax: (204) 983-7557 If you hold a UNIX shell up to your ear, can you hear the C? - Robin [EMAIL PROTECTED] 20-Dec-01 8:20:26 AM I don't see how aggregate rate could exceed network rate. Aggregate is Bytes Transferred divided by Elapsed Time, and Network is Bytes Transferred divided by Data Transfer Time what were those values in the session where Aggregate was greater than Network? How could Data Transfer Time be greater than Elapsed Time? Must be a bug! Robin Sharpe Berlex Labs Miles Purdy PURDYM@FIPD. GC.CATo:[EMAIL
Re: ATL P2000 dlt 7000 question
Paul, I remember this same topic being discussed several months ago... I thought it was impossible... how can the library move the slider on the cartridge!? We have a P3000 and have never seen this happen. I don't remember the resolution from that last discussion (or if there was one)... but I'll say again what I said then... call ATL tech support! They are usually pretty helpful. Robin Sharpe Berlex Labs Coviello, Paul PCoviello@CM To:[EMAIL PROTECTED] C-NH.ORG cc:(bcc: Robin Sharpe/WA/USR/SHG) Subject: 12/24/01 ATL P2000 dlt 7000 question 12:40 PM Please respond to ADSM: Dist Stor Manager Hi we have an ATL P2000 with 4 dlt7000's 100 tape capacity. the operations group is seeing DLT tapes coming out of the library write protected. Would anyone care to guess at this? because I can't even come up with a rational explanation nor logical one. :-) I have asked if it is the same tapes, ( going thru list this week) thinking they might have a bad switch but then what moves it! thanks Paul Paul J Coviello Sr Systems Analyst Catholic Medical Center 2456 Brown Ave Manchester NH 03103 (603) 663-5326 [EMAIL PROTECTED]
Re: Please Explain
I don't see how aggregate rate could exceed network rate. Aggregate is Bytes Transferred divided by Elapsed Time, and Network is Bytes Transferred divided by Data Transfer Time what were those values in the session where Aggregate was greater than Network? How could Data Transfer Time be greater than Elapsed Time? Must be a bug! Robin Sharpe Berlex Labs Miles Purdy PURDYM@FIPD. GC.CATo:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 12/20/01 Subject: 08:44 AM Re: Please Explain Please respond to ADSM: Dist Stor Manager The network transfer rate is the time is takes to send the file to be backed up to the TSM server. The aggregate is the total KB backed up / the total time. The difference is the processing time. The time to contact the TSM server and check if the file needs to be backed up. Interestingly enough, my aggregate usually exceeds the network, I was asking yesterday if any one else sees this. Miles -- Miles Purdy System Manager Farm Income Programs Directorate Winnipeg, MB, CA [EMAIL PROTECTED] ph: (204) 984-1602 fax: (204) 983-7557 If you hold a UNIX shell up to your ear, can you hear the C? - [EMAIL PROTECTED] 20-Dec-01 4:45:16 AM Hi I run a full backup of a Netware 5 with Tivoli client 4.2.0 here the statistics: 12/18/2001 13:58:12 Total number of objects inspected: 33,310 12/18/2001 13:58:12 Total number of objects backed up: 33,104 12/18/2001 13:58:12 Total number of objects updated: 0 12/18/2001 13:58:12 Total number of objects rebound: 0 12/18/2001 13:58:12 Total number of objects deleted: 0 12/18/2001 13:58:12 Total number of objects expired: 0 12/18/2001 13:58:12 Total number of objects failed: 3 12/18/2001 13:58:12 Total number of bytes transferred: 1.33 GB 12/18/2001 13:58:12 Data transfer time: 138.38 sec 12/18/2001 13:58:12 Network data transfer rate:10,099.94 KB/sec 12/18/2001 13:58:12 Aggregate data transfer rate:802.25 KB/sec 12/18/2001 13:58:12 Objects compressed by:0% 12/18/2001 13:58:12 Elapsed processing time: 00:29:02 Can anybody explain why my Network data transfer rate is so high but my Aggregate data transfer rate is low T.I.A Regards Robert Ouzen E-mail: [EMAIL PROTECTED]
Re: Incremental forever -- any problems? (Scary thoughts)
Oops - Forwarded by Robin Sharpe/WA/USR/SHG on 12/20/01 10:01 AM - Robin Sharpe 12/20/01 09:14 To: [EMAIL PROTECTED] AM cc: Subject: Re: Incremental forever -- any problems? (Scary thoughts)(Document link: Robin Sharpe) But wait a miniute doesn't the multi streamed backup work by allocating a session to each filespace (filsystem or disk partition)? I'm not sure you would get multiple streams if you are only backing up a single filespace. If that is true, then I would guess that Tivoli will implement a multi-stream restore in the same way -- one session for each filespace being restored. On the other hand, I don't think there is any technical reason why the backup couldn't split even a single directory between two sessions. But that's not true for restores (as you have all said already)... it will only work if the data is on multiple tapes. However, let's assume we have a non-collocated pool containing a client with a wide variety of data... 30% that rarely changes, 30% changes weekly, and 40% changes daily, and all of the data is on a single filespace. After several months of backups the active backup data will probably be spread over several tapes. We now need to restore that filespace due to a disk crash. A well-implemented multi-stream restore should be able to sort the data by tape volume and start a session for each volume. Note that we can't even do that manually now, because we have no way of knowing what tape contains what active versions. I think it's important for Tivoli to implement multi streaming in this way since so many NT clients have only one or two filespaces... In our shop, most have two -- a C: drive for NT and a D: drive for files or Notes or whatever. Robin Sharpe Berlex Labs Michael Bartl michael.bart [EMAIL PROTECTED] To:[EMAIL PROTECTED] cc:(bcc: Robin Sharpe/WA/USR/SHG) 12/20/01 Subject: 03:51 AM Re: Incremental forever -- any problems? (Scary thoughts) Please respond to michael.bartl Hi Kelly, your're right, a multistream restore is one of the items on top of our wishlist. Just one thought on collocation: With collocation you can't mount as many tapes as without, ok. But with collocation, data from one node is much more compactly written to one tape. So when you don't use collocation and you have to mount 50 instead of 7 tapes (or even 7 instead of one) you have to keep in mind that TSM has to skip 6/7 of the data on the tapes while reading them. So collocation would remain an important feature, the benefit just decreasing in size (compared to now). Best regards, Michael Kelly Lipp wrote: On the wishlist item: I have heard talk about implementing an automatic multi-stream restore similar to what we have now with backup and resourceutilization. This will be slick but then only really slick if one uses collocate=filespace. I believe it will still provide better restore times since it will know what tape volumes client data is on and will mount multiple volumes. If one thinks about this for a few minutes, one realizes the whole thing is damn complicated. Also, it would appear that collocation would not be a good thing if you use this technique: you want data on multiple tapes. -- Michael Bartl mailto:[EMAIL PROTECTED] Office of Technology, IT Germany/AustriaTel: +49-89-92699-806 Cable Wireless Deutschland GmbH. Fax: +49-89-92699-302 Landsberger Str. 155, D-80687 Muenchen http://www.cw.com/de