Re: TSM Operational Reporting notification of Failed/Missed schedules
I ended up creating a user group in our email system, and adding the individuals I want to receive this email to that group. That should work (I will test it out today), but I image the %1 field will be replaced with the group contact name, instead of an individual's name. Still asking the question... does TOR allow multiple contacts and email addresses? If so, what is the appropriate syntax? From: Todd Lundstedt Sent: Wednesday, September 21, 2005 2:47 PM To: ADSM-L@VM.MARIST.EDU Subject: TSM Operational Reporting notification of Failed/Missed schedules In the properties box of the daily report, on the automatic notification tab, after I query the server for the list of nodes and contacts, one of my nodes doesn't show up with the correct contact information. That node happens to have multiple contacts; four different contacts. I remember entering them in seperated by commas without spaces for the contact field and for the email address field. I checked the failed and missed check boxes, and performed the update. The contact info on the TSM server for that node now shows [EMAIL PROTECTED],[EMAIL PROTECTED],[EMAIL PROTECTED],[EMAIL PROTECTED],contact1,contact 2,contact3,contact4,3 I believe the 3 at the end indicates both failed and missed was selected. When I query the server after the update, the contact box is populated with [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] , and the email field is populated with [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] . The failed and missed check boxes are no longer checked. Obviously, I have guessed an incorrect syntax for inputting multiple contacts in those fields, if it can even be done. Are multiple notification contacts allowed with TOR, and if so, what is the correct syntax?
TSM Operational Reporting notification of Failed/Missed schedules
In the properties box of the daily report, on the automatic notification tab, after I query the server for the list of nodes and contacts, one of my nodes doesn't show up with the correct contact information. That node happens to have multiple contacts; four different contacts. I remember entering them in seperated by commas without spaces for the contact field and for the email address field. I checked the failed and missed check boxes, and performed the update. The contact info on the TSM server for that node now shows [EMAIL PROTECTED],[EMAIL PROTECTED],[EMAIL PROTECTED],[EMAIL PROTECTED],contact1,contact 2,contact3,contact4,3 I believe the 3 at the end indicates both failed and missed was selected. When I query the server after the update, the contact box is populated with [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] , and the email field is populated with [EMAIL PROTECTED] mailto:[EMAIL PROTECTED] . The failed and missed check boxes are no longer checked. Obviously, I have guessed an incorrect syntax for inputting multiple contacts in those fields, if it can even be done. Are multiple notification contacts allowed with TOR, and if so, what is the correct syntax?
Re: TSM Server Recovery (AIX)
Stef, We also perform backups to a local 20GB 4mm tape. Our TSM server is a B80. That mksysb tape will not restore to a p650. From what I have been told and what I have read, I need to have a set of install CDs at the same OS-ML as the mksysb. Since I only have install CDs for 5200-01 and my mksysb is at 5200-03, I was unable to restore my TSM server from that mksysb until I found a spare B80 lying around (thank goodness our DR site vendor had one at that time, but there is no guarantee they will have one next time, or when we really need it) -- it's a long story as to how this happened. To be better prepared for every eventuality, I need to find a way to create install CDs at any OS-ML currently in use at our shop. Aaron (from a different reply), I have messed around with NIM mksysbs and the mkcd command quite a bit. Our TSM server is also our NIM server. Currently, it is only setup with an NFS exported mountpoint to which other AIX servers perform their mksysbs. Using the TSM server alone, I have not yet been able to create a mksysb CD or DVD image of another system that will boot and restore that system's rootvg. Realize, I still need to be able to restore those client mksysb images to potentially different hardware, as well, so I still need to have a way to create a bootable install CD image of the existing OS-ML. I hope that better explains what I am looking for. Thanks, Todd -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Stef Coene Sent: Monday, September 19, 2005 1:25 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] TSM Server Recovery (AIX) On Monday 19 September 2005 19:47, Todd Lundstedt wrote: This is more along the lines of AIX server recovery, but it plays into TSM, too. Without using addtional software, just AIX, how does one create a boot CD that contains all the necessary info to restore a server using a mksysb tape? I understand the process for recovering an AIX server from a mksysb tape backup to different hardware would be to boot the server from a CD of the same OS-ML as the mksysb backup. My TSM server is running on AIX 5200-03. The install media I have is 5200-01. How do I create a boot CD image (ISO, so I can burn the ISO image using a PC CD-W) that contains the 5200-03 files so I can recover the TSM server? My current working documentation requires an install of 5200-01, patch to -03, and install TSM and then restore the DB. That works OK, but takes additional time, and still won't get back everything I need to allow some other node restores to start. I will still have to rebuild some things from scratch. We always backup the TSM server (AIX) on the local tape streamer. So we have a bootable mksysb tape that also contains the TSM db backup. For the other AIX servers we use nim to make backups. Nim allows you to make a mksysb to a nim master and stores the mksysb as a file on the nim master (this is also the TSM server). The AIX servers can boot from the nim master over the network and restore the mksysb. Stef
Re: TSM Operational Reporting just stops functioning
Bingo! Mike and Steve hit it. The password thing was not it, since it does work for a while. After reading about IC44976, I found the trace file and could see those error messages in the file. I am off to find the latest patch for TOR 5.3. Unfortunately, the files at ftp://ftp.software.ibm.com/storage/tivoli-storage-management/techprev/ts mopreport/latest have not been updated since 2003, so I doubt this is really the latest available. Does anyone have a better source for TOR latest code? Until then, I will be scheduling a daily restart of that service. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Schaub, Steve Sent: Thursday, September 15, 2005 5:27 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] TSM Operational Reporting just stops functioning I would suspect IC44976, since Todd noted that TOR would work for a day or two before entering the Twilight Zone. I have also been bitten by the expired admin pwd bug, but that stopped TOR cold in my case, so I didn't think it fit this particular problem. -steve -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of E Mike Collins Sent: Thursday, September 15, 2005 12:57 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] TSM Operational Reporting just stops functioning Hi Todd, Steve, All, Two likely causes are documented in APARs IC43649, and IC44976. You can go to ibm.com and search for those for more information. In a nut shell, IC43649 documents a restriction for TOR where it doesn't surface an expired admin password for the account it uses to communicate with the server. An expired password can cause all TOR worker threads to be consumed resulting in no further scheduled reports being sent out. The recommendation is to reset any expired passwords and then set the password to not expire. IC44976 is fixed in 5.2.6 and will be available in 5.3.2 when that ships. It fixes a resource leak that will also cause this behavior. If you need additional information please send me a note directly. Best Regards, Mike Collins, [EMAIL PROTECTED] Ref: Subject: TSM Operational Reporting just stops functioning TOR is 5.3.1.0, running on W2K3. The hardware also runs ISC/AC, but we don't use that much. TOR is set to monitor two TSM servers, one at 5.3, and one at 5.2, both running AIX, if it matters. For each TSM server, I have the standard hourly report, and two daily reports; the standard daily report split up into two - a summary report and a detail report. By looking at the current reports in MMC, it appears that TOR simply stopped running the reports. The tsmrept service is still running. There is nothing in the application or system event logs to indicate an issue. Looking at the date/time on the last reports run, all of the daily reports appeared to run at the prescribed time for that day. The hourly reports ended up running for their last time at different times (TSM server1's last hourly report was several hours later than TSM server2's last hourly report). Once the hourly reports stop running, subsequent daily reports will not run, either. Stopping and starting the tsmrept service fixes this for a while. Sometimes it will work fine for only a day, sometimes for three days. I have verified that the select commands are not getting to the TSM server: again, the service is still running. I deactived the detail level report, but that hasn't helped anything. I found a file tecinfo.txt in the Console\TEC folder of TSM. The log information there supports my observations; hourly reports running fine, daily reports running fine, then hourlys stopping for one server, and then stopping for the other server, and no additional reports after that, until I stop and start the service. Any thoughts? Todd Please see the following link for the BlueCross BlueShield of Tennessee E-mail disclaimer: http://www.bcbst.com/email_disclaimer.shtm
Re: TSM Operational Reporting just stops functioning
Aah.. I get it, now. The SRV file being the Server binaries and stuff, and CON being the management console; silly me, looking for TOR or something like that. Mike, Any idea when 5.3.2 will be out? And are there any known incompatibilities between the 5.3.2 TOR and the 5.2.2.5 AIX servers or 5.3.1.0 AIX servers? In that respect, because 5.2.6.1 should already have the fix in place, shouldn't I be able to remove my 5.3 TOR install, and go with the 5.2.6.1 TOR? And will the 5.2.6.1 TOR work with a 5.3 server? If so, are there clear instructions around for downleveling TOR? Thanks again for all the help, gang! Todd -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of William Boyer Sent: Thursday, September 15, 2005 9:53 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] TSM Operational Reporting just stops functioning It's in ftp://ftp.software.ibm.com/storage/tivoli-storage-management/patches/ser ver/NT/5.3.1.5/TSMCON5315_WIN.exe Bill Boyer Some days you're the bug, some days you're the windshield - ?? -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Todd Lundstedt Sent: Thursday, September 15, 2005 10:40 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: TSM Operational Reporting just stops functioning Bingo! Mike and Steve hit it. The password thing was not it, since it does work for a while. After reading about IC44976, I found the trace file and could see those error messages in the file. I am off to find the latest patch for TOR 5.3. Unfortunately, the files at ftp://ftp.software.ibm.com/storage/tivoli-storage-management/techprev/ts mopreport/latest have not been updated since 2003, so I doubt this is really the latest available. Does anyone have a better source for TOR latest code? Until then, I will be scheduling a daily restart of that service. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Schaub, Steve Sent: Thursday, September 15, 2005 5:27 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] TSM Operational Reporting just stops functioning I would suspect IC44976, since Todd noted that TOR would work for a day or two before entering the Twilight Zone. I have also been bitten by the expired admin pwd bug, but that stopped TOR cold in my case, so I didn't think it fit this particular problem. -steve -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of E Mike Collins Sent: Thursday, September 15, 2005 12:57 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] TSM Operational Reporting just stops functioning Hi Todd, Steve, All, Two likely causes are documented in APARs IC43649, and IC44976. You can go to ibm.com and search for those for more information. In a nut shell, IC43649 documents a restriction for TOR where it doesn't surface an expired admin password for the account it uses to communicate with the server. An expired password can cause all TOR worker threads to be consumed resulting in no further scheduled reports being sent out. The recommendation is to reset any expired passwords and then set the password to not expire. IC44976 is fixed in 5.2.6 and will be available in 5.3.2 when that ships. It fixes a resource leak that will also cause this behavior. If you need additional information please send me a note directly. Best Regards, Mike Collins, [EMAIL PROTECTED] Ref: Subject: TSM Operational Reporting just stops functioning TOR is 5.3.1.0, running on W2K3. The hardware also runs ISC/AC, but we don't use that much. TOR is set to monitor two TSM servers, one at 5.3, and one at 5.2, both running AIX, if it matters. For each TSM server, I have the standard hourly report, and two daily reports; the standard daily report split up into two - a summary report and a detail report. By looking at the current reports in MMC, it appears that TOR simply stopped running the reports. The tsmrept service is still running. There is nothing in the application or system event logs to indicate an issue. Looking at the date/time on the last reports run, all of the daily reports appeared to run at the prescribed time for that day. The hourly reports ended up running for their last time at different times (TSM server1's last hourly report was several hours later than TSM server2's last hourly report). Once the hourly reports stop running, subsequent daily reports will not run, either. Stopping and starting the tsmrept service fixes this for a while. Sometimes it will work fine for only a day, sometimes for three days. I have verified that the select commands are not getting to the TSM server: again, the service is still running. I deactived the detail level report, but that hasn't helped anything. I found a file tecinfo.txt in the Console\TEC folder of TSM. The log information there supports my observations; hourly reports running fine, daily reports running fine, then hourlys stopping for one server
Re: backup stgpool
Sandra, As I re-read through this, I realize I made a mistake. Using delete volume will not adhere to the reuse delay. Reclamation set to 1 will. But I also think there are other things that should be considered, most of which were addressed by Andy in other posts. Sorry if my mistake cause you any issues. -Original Message- From: Todd Lundstedt Sent: Thursday, September 08, 2005 9:03 AM To: 'ADSM-L@VM.MARIST.EDU' Subject: RE: [ADSM-L] backup stgpool Sandra, As I read through your initial request, it sounds like, every Friday, you want all of your offsite copy pool tapes consolidated onto fewer tapes. If that is what you are wanting, you can delete vol with discarddata=yes for each of your offsite pool tapes prior to your backup stgpool (now that you have the reuse set to 4 days, they still won't be eligible for return for another four days). Or, you can set your reclamation=1 on the offsite copy storage pool (I am not sure you can set it to 0, I have never tried). -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Sandra Sent: Thursday, September 08, 2005 1:21 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] backup stgpool Hmmm thanks andy, I have changed the reuse delay to 4, that is the same as dbbackup expiration days. Ok..i have to have two copy storage pools both having a complete and separate set of tapes. From this i understand that everytime the stgpool backup is performed, there is no way to tell TSM that you need to take full backup of primary pool instead of incremental from what it did last day. Kind Regards Sandra [EMAIL PROTECTED] wrote : Hi Sandra, I'm sure you'll get plenty of advice and suggestions from others, but here are my two cents: Once concern right off the bat is the reuse delay of zero, which puts at risk your goal of 100% restore capability of up to 4 days ago. Instead, your reuse delay should be the same as your dbbackup and plan file expiration settings. Consider the case where on day 1, tape 'A' contains valid data that is expired on day 2. Tape 'A' is now empty, so it is available for immediate reuse. On day 3, the tape is written over with new data. Now it is day 4, and you need to restore your database back to day 1. You restore the database back to day 1. Now you want to restore data from tape 'A' that was there on day 1... but that data no longer exists because it was written over with new data on day 3... and your 100% restore capability is out the window. From what you wrote, I am under the impression that you think you need multiple sets of copy storage pools to allow you to restore up to 4 days ago. But this is not true. Since each copy pool is a logical mirror of the primary pools, a single copy pool is all that is necessary to restore your data up to four days ago (assuming everything else is also configured to minimally meet the four day requirement). If you want an onsite copy pool and an offsite copy pool, then have two copy pools: one that stays onsite, and another that goes offsite. You can then run BACKUP STGPOOL multiple times to back up the primary pools to each of your copy pools. If you need more onsite copy pools, you can define them and back up the storage pools to them. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-09-07 21:41:21: Dear All, I am running TSM 5.2.3 on Windows 2003 with 3582 Library with 2 drives. Ultrium2 tapes with 200GB space. Daily incremental backup of about 5GB. I have setup a copypool to which i backup my primary pool daily. The plan file (after running prepare) has ever increasing tapes required for primary stgpool restore. My question: Can i take full backup of stgpool daily and not the incremental on daily basis. The command refernce says it won't copy any file to copypool until the file is damaged or not available in copypool... that is sort of incremental. Purpose is to send one set of DR tapes to offsite weekly on friday. I want this set to be 100% capable of restoring my environment while at the same time, having 3 sets of DR at main site so that I can revert to a maximum of 4 days. DBbackup and plan file expiration is 4 days, and volume reuse delay in all stgpools is 0. I want to reuse the tapes immediately. Can anyone enlighten me on this. Regards, Sandra _ Sent via SUPERwebmail - Supernet web-based email service http://www.super.net.pk/mail _ Sent via SUPERwebmail - Supernet web
TSM Operational Reporting just stops functioning
TOR is 5.3.1.0, running on W2K3. The hardware also runs ISC/AC, but we don't use that much. TOR is set to monitor two TSM servers, one at 5.3, and one at 5.2, both running AIX, if it matters. For each TSM server, I have the standard hourly report, and two daily reports; the standard daily report split up into two - a summary report and a detail report. By looking at the current reports in MMC, it appears that TOR simply stopped running the reports. The tsmrept service is still running. There is nothing in the application or system event logs to indicate an issue. Looking at the date/time on the last reports run, all of the daily reports appeared to run at the prescribed time for that day. The hourly reports ended up running for their last time at different times (TSM server1's last hourly report was several hours later than TSM server2's last hourly report). Once the hourly reports stop running, subsequent daily reports will not run, either. Stopping and starting the tsmrept service fixes this for a while. Sometimes it will work fine for only a day, sometimes for three days. I have verified that the select commands are not getting to the TSM server: again, the service is still running. I deactived the detail level report, but that hasn't helped anything. I found a file tecinfo.txt in the Console\TEC folder of TSM. The log information there supports my observations; hourly reports running fine, daily reports running fine, then hourlys stopping for one server, and then stopping for the other server, and no additional reports after that, until I stop and start the service. Any thoughts? Todd
Re: TSM for System Backup and Recovery restore issues
Thanks, David. As it turns out, a Sysback developer saw my original note, and sent me an email -- wow!! how did I rate that!? Anyway, he had me upgrade to a newer version, and after talking to him some more about my setup, I realized I didn't have all the required software installed into the appropriate places, and we corrected that, as well. One thing he was concerned about was the fact that the Installation Options menu had no options at all, when it should have at least had a CD and a network entry, even though I didn't have all of the required software components installed in the right places. Further attempts still failed. As we continued to look at the issues, both of us very perplexed, I noticed something that I had not realized was happening... I was receiving several errors about no space when the SBR Installation and Maintenance menu was loading (it happened so fast, I couldn't really tell what was going on until I scrolled up). Then, he had me check the size of the ramfs, and it was waaay to small. He ended up walking me through some modifications of a couple of scripts to increase the size of the ramfs, and then we rebuilt the boot image. This time it worked, and I have been successful restoring the backup image I created. He will be looking into the ramfs issue, and I am back to testing this product out. Thanks for the note, Todd -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of David E Ehresman Sent: Thursday, September 08, 2005 2:54 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] TSM for System Backup and Recovery restore issues If you're still having problems, you might call tivoli. There was a min level of sysback that I needed for aix 5.2 and I think it might have been 5.6.6 but I'm not sure. [EMAIL PROTECTED] 09/07/05 5:53 PM Server TSM 5.2.2.5 running on AIX 5.2 NIM master TSM4SBR 5.6.5.0 Client AIX 5.2 TSM Client 5.2.2.14 (or whatever is fairly recent in the 5.2.2.x range) TSM4SBR 5.6.5.0 I have been successful creating the TSM4SBR System Image backup of the Client to the virtual TSM device, and verified that the backups exist as a /TSM4SBR filespace in the Server's storage pools. I have installed TSM4SBR into the NIM SPOT assigned to this Client, rebuilt the Network Boot Image, and reallocated the SPOT and LPP_Resource to the Client via the NIM Master's SMIT panels. I have been successful booting the client from the network, pointing the client to the NIM Master's IP address as a bootp server. I have received the TSM for System Backup and Recovery Installation and Maintenance screen after boot. The Installation Device is [not set]. When I use the menu option to Change Installation Device, I receive the next panel that is mostly blank, and indicates There are no tape drives or network devices available to install from. (bad grammar, IBM). Examples of this panel in the manual indicate there might be CD, Tape, and a tsmdev device listed. Nothing is listed. What am I missing? Where do I set or define the installation devices available to a boot image. I assume I will have to rebuild the boot image once I make that definition. Using the NIM Master's smit panels, I have tried entering network identifiers (ent1, etc) into the Software Install Device field of different Install Client Defaults panels, but that has not helped. Any help would be appreciated. If you think you could help better via a phone call, I can place that call. Email me directly with your phone number. Thanks in advance, Todd Lundstedt Via Christi Regional Medical Center Wichita KS
Re: backup stgpool
Sandra, As I read through your initial request, it sounds like, every Friday, you want all of your offsite copy pool tapes consolidated onto fewer tapes. If that is what you are wanting, you can delete vol with discarddata=yes for each of your offsite pool tapes prior to your backup stgpool (now that you have the reuse set to 4 days, they still won't be eligible for return for another four days). Or, you can set your reclamation=1 on the offsite copy storage pool (I am not sure you can set it to 0, I have never tried). -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Sandra Sent: Thursday, September 08, 2005 1:21 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] backup stgpool Hmmm thanks andy, I have changed the reuse delay to 4, that is the same as dbbackup expiration days. Ok..i have to have two copy storage pools both having a complete and separate set of tapes. From this i understand that everytime the stgpool backup is performed, there is no way to tell TSM that you need to take full backup of primary pool instead of incremental from what it did last day. Kind Regards Sandra [EMAIL PROTECTED] wrote : Hi Sandra, I'm sure you'll get plenty of advice and suggestions from others, but here are my two cents: Once concern right off the bat is the reuse delay of zero, which puts at risk your goal of 100% restore capability of up to 4 days ago. Instead, your reuse delay should be the same as your dbbackup and plan file expiration settings. Consider the case where on day 1, tape 'A' contains valid data that is expired on day 2. Tape 'A' is now empty, so it is available for immediate reuse. On day 3, the tape is written over with new data. Now it is day 4, and you need to restore your database back to day 1. You restore the database back to day 1. Now you want to restore data from tape 'A' that was there on day 1... but that data no longer exists because it was written over with new data on day 3... and your 100% restore capability is out the window. From what you wrote, I am under the impression that you think you need multiple sets of copy storage pools to allow you to restore up to 4 days ago. But this is not true. Since each copy pool is a logical mirror of the primary pools, a single copy pool is all that is necessary to restore your data up to four days ago (assuming everything else is also configured to minimally meet the four day requirement). If you want an onsite copy pool and an offsite copy pool, then have two copy pools: one that stays onsite, and another that goes offsite. You can then run BACKUP STGPOOL multiple times to back up the primary pools to each of your copy pools. If you need more onsite copy pools, you can define them and back up the storage pools to them. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-09-07 21:41:21: Dear All, I am running TSM 5.2.3 on Windows 2003 with 3582 Library with 2 drives. Ultrium2 tapes with 200GB space. Daily incremental backup of about 5GB. I have setup a copypool to which i backup my primary pool daily. The plan file (after running prepare) has ever increasing tapes required for primary stgpool restore. My question: Can i take full backup of stgpool daily and not the incremental on daily basis. The command refernce says it won't copy any file to copypool until the file is damaged or not available in copypool... that is sort of incremental. Purpose is to send one set of DR tapes to offsite weekly on friday. I want this set to be 100% capable of restoring my environment while at the same time, having 3 sets of DR at main site so that I can revert to a maximum of 4 days. DBbackup and plan file expiration is 4 days, and volume reuse delay in all stgpools is 0. I want to reuse the tapes immediately. Can anyone enlighten me on this. Regards, Sandra _ Sent via SUPERwebmail - Supernet web-based email service http://www.super.net.pk/mail _ Sent via SUPERwebmail - Supernet web-based email service http://www.super.net.pk/mail
TSM for System Backup and Recovery restore issues
Server TSM 5.2.2.5 running on AIX 5.2 NIM master TSM4SBR 5.6.5.0 Client AIX 5.2 TSM Client 5.2.2.14 (or whatever is fairly recent in the 5.2.2.x range) TSM4SBR 5.6.5.0 I have been successful creating the TSM4SBR System Image backup of the Client to the virtual TSM device, and verified that the backups exist as a /TSM4SBR filespace in the Server's storage pools. I have installed TSM4SBR into the NIM SPOT assigned to this Client, rebuilt the Network Boot Image, and reallocated the SPOT and LPP_Resource to the Client via the NIM Master's SMIT panels. I have been successful booting the client from the network, pointing the client to the NIM Master's IP address as a bootp server. I have received the TSM for System Backup and Recovery Installation and Maintenance screen after boot. The Installation Device is [not set]. When I use the menu option to Change Installation Device, I receive the next panel that is mostly blank, and indicates There are no tape drives or network devices available to install from. (bad grammar, IBM). Examples of this panel in the manual indicate there might be CD, Tape, and a tsmdev device listed. Nothing is listed. What am I missing? Where do I set or define the installation devices available to a boot image. I assume I will have to rebuild the boot image once I make that definition. Using the NIM Master's smit panels, I have tried entering network identifiers (ent1, etc) into the Software Install Device field of different Install Client Defaults panels, but that has not helped. Any help would be appreciated. If you think you could help better via a phone call, I can place that call. Email me directly with your phone number. Thanks in advance, Todd Lundstedt Via Christi Regional Medical Center Wichita KS
Image Backups and SANs
TSM server 5.2.2.5 on AIX 5.2 BAClient 5.2.2.0 on Windows 2003 I noticed the following in the help regarding BACKUP IMAGE: Image backup of IBM TotalStorage SAN File Systems is not supported. ... and in the Backup/Archive Installation manual, 5.2.2, Technical changes for version 5.2.2 - Dec 2003 section Backup and restore support for IBM TotalStorage SAN File systems Backup and restore for IBM TotalStorage SAN File Systems is supported on the Windows 2000 client. See Supported files systems ... and that section also indicates IBM TotalStorage SAN File System is supported on Windows 2000 only. Is data on an IBM Shark considered on an IBM TotalStorage SAN File System? Does the support also extend to Windows 2003? Thanks in advance
migr_mb in stgpools is not reporting correctly
TSM 5.2.2.5 on AIX 5200-03 select stgpool_name, migr_mb from stgpools Why do all but one (sometimes two) of my disk storage pools that migrate to tape only have 0.0? It even displays this value when a migration is in progress, with GBs of data transferred. Am I missing something here? The two that do have values are less than 1,000 MB (one around 40 MB, the other around 720 MB).
Re: Server 5.2 with 5.1 clients?
The only issue we have seen here is NT4 is no longer supported with 5.2 Storage Agent. That hit us hard since we weren't informed of it by the company that performed the 4.2 - 5.2 upgrade (with an AIX upgrade as well). We had to scramble to re-schedule large database servers so all the databases weren't backing up FULL every day. Other than that, the older B/A clients appear to work well for us (so far, we haven't had to do a restore of an NT4 node from our 5.2 server, yet). -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Ben Bullock Sent: Wednesday, March 02, 2005 12:18 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Server 5.2 with 5.1 clients? We have many old clients backing up to our 5.2 TSM servers, a few TSM V3 clients, and they still work great. IBM does not guarantee that they will work, and we would probably have issues getting support, but we find in most cases they continues to work... Until they don't :-) 5.1.5 clients should be fine, IMHO. Ben -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Rob Hefty Sent: Wednesday, March 02, 2005 11:13 AM To: ADSM-L@VM.MARIST.EDU Subject: Server 5.2 with 5.1 clients? Hello all, We have a TSM server version 5.2.3.2 and have been upgrading clients to be on the 5.2 client level and found that our old WinNT 4.0 servers are not compatible with this version level. Will these older clients (version 5.1.5.0) be able to backup\restore and archive\retrieve to a 5.2 server properly? If so, is this something that is guaranteed by IBM, or merely a loophole that works? Thanks, Rob
Re: Select for Tape Storage Pool Report
Thanks, Here is the script in it's final form if any one else wants it. Adjust your devclass and char/decimal output and column titles as needed.. select - cast(a.stgpool_name as char(22)) as Stg Pool , - cast((a.est_capacity_mb/1024/1024) as dec(5,2)) as TB, - a.pct_utilized as PctUtl, - a.pct_logical as Logi, - cast(a.recl_running as char(4)) as Run?, - cast(a.reclaim as dec(3)) as Recl, - cast(a.maxscratch as dec(3)) as Max, - cast(count(*) as dec(3)) as VolUsed, - cast(a.maxscratch - count(*) as dec(3)) as Diff - from - stgpools a, - volumes b - where - devclass in ('LTO','LTO2') and - a.stgpool_name=b.stgpool_name - group by - a.stgpool_name, - a.est_capacity_mb, - a.pct_utilized, - a.pct_logical, - a.recl_running, - a.reclaim, - a.maxscratch -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Robert Ouzen Sent: Friday, February 18, 2005 11:44 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Select for Tape Storage Pool Report Hi Todd This Script will give you the number of volumes per stg and maxscratch allocated select a.stgpool_name,a.maxscratch,count(*) as Number of Vols from stgpools a, volumes b where a.stgpool_name = b.stgpool_name and a.devclass = 'SCALARCLASS' group by a.stgpool_name,a.maxscratch Regards Robert Ouzen Haifa University Israel -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Todd Lundstedt Sent: Saturday, February 19, 2005 12:00 AM To: ADSM-L@VM.MARIST.EDU Subject: Select for Tape Storage Pool Report I created this SQL select statement... select - cast(stgpool_name as char(22)) as Stg Pool , - cast((est_capacity_mb/1024/1024) as decimal(5,2)) as TB, - pct_utilized as Util, - pct_logical as Logi, - cast(recl_running as char(4)) as Run?, - cast(reclaim as dec(3)) as Recl, - cast(maxscratch as dec(3)) as Max - from stgpools - where devclass in ('LTO','LTO2') It outputs exactly what I think it should... Stg Pool TB Util LogiRun? Recl Max ----------- L1_CPY_DBVL_LTO2_OFF1 19.06 14.5 100.0NO100 55 L1_CPY_DB_LTO_OFF1 18.18 0.8 100.0NO100 100 L1_CPY_DB_LTO_ON 0.00 0.0 100.0NO100 25 L1_CPY_DSKIMG_LTO2_OFF 0.00 0.0 100.0NO100 30 L1_CPY_LTO2_OFF110.91 3.9 99.9NO100 30 L1_CPY_LTO_OFF1 9.50 35.9 99.5YES60 50 L1_CPY_LTO_ON0.00 0.0 100.0NO100 20 L1_PRI_DBVL_LTO216.65 16.6 100.0NO100 40 L1_PRI_DB_LTO5.43 2.7 100.0NO100 24 L1_PRI_DSKIMG_LTO2 0.00 0.0 100.0NO 60 10 L1_PRI_LTO 5.85 58.5 99.1NO100 29 L1_PRI_LTO2 3.00 14.4 99.9NO1008 L2_CPY_DB_LTO_OFF1 0.00 0.0 100.0NO100 20 L2_CPY_DSKIMG_LTO2_OFF 0.00 0.0 100.0NO100 30 L2_CPY_LTO_OFF1 5.07 22.2 99.8NO 60 30 L2_PRI_DB_LTO0.00 0.0 100.0NO1001 L2_PRI_DSKIMG_LTO2 0.00 0.0 100.0NO 60 10 L2_PRI_LTO 2.32 48.8 99.6NO100 15 L3_CPY_DB_LTO_OFF1 3.69 15.7 100.0YES60 20 L3_CPY_DOM_LTO_OFF1 4.54 11.1 100.0YES60 25 L3_CPY_DSKIMG_LTO2_OFF 0.00 0.0 100.0NO100 30 L3_CPY_LTO_OFF1 6.25 61.0 99.8NO 60 35 L3_CPY_MAIL_LTO_OFF1 3.56 5.7 100.0NO100 20 L3_PRI_DB_LTO2.64 22.1 100.0NO100 10 L3_PRI_DOM_LTO 1.74 29.0 100.0NO100 10 L3_PRI_DSKIMG_LTO2 0.00 0.0 100.0NO 60 10 L3_PRI_LTO 5.68 67.3 99.6NO100 30 L3_PRI_MAIL_LTO 1.77 17.0 100.0NO100 10 LTOTAPE 11.54 25.7 99.7NO 60 50 LTOTAPE_SQL 1.81 3.6 100.0NO100 10 OFFSITE 19.16 15.8 99.9YES60 100 SPACEMGPOOL 0.00 0.0 100.0NO 600 But, I want to add yet one more column.. the count of volumes currently assigned to that storage pool. I can find that information with the following select statement... select - count(*) - from volumes - where
Why can't I delete an other server entry?
TSM 5.2.2.5 on AIX 5.2 ML3 I installed and setup Storage Agent on a W2K node for testing purposes. Everything tested out fine, and now I am wanting to remove everything, including the other server entry on the TSM server. I have already uninstalled the Storage Agent and device drivers from the W2K node. I have deleted all the filespaces related to that node name (a special node name setup to test the storage agent, the original node and filespaces still exist, but it never used the storage agent to do backups). I did all that last week sometime. Today, when I try to delete the other server entry for the Storage Agent, it tells me the server is currently in use. I deleted the nodename that we used for testing, and ran audit lic... no change. Nothing is listed in q ses related to this node. I checked netstat to verify there were no connections to that IP address... What am I missing? (I don't know if it matters, but we were testing LVSA and Image backups using the storage agent). TIA Todd
Re: Why can't I delete an other server entry?
Sorry.. Forgot to mention what I had checked out as a result of the ANR1664W message... I have no devtype definitions of server (at least, in the GUI, when I go to Objects, Server Storage, Device Classes, Server Device Classes, there are no entries)... We don't have an event server... And I addressed the connection issues with comments re: sessions and netstat connections.. Still can't figure it out.. Todd -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Andrew Raibeck Sent: Monday, February 21, 2005 1:26 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Why can't I delete an other server entry? Check the DELETE SERVER syntax, as it includes a list of items that will cause the command to fail. In particular, make sure you don't have any device SERVER-type classes that point to your other server. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-02-21 12:20:49: TSM 5.2.2.5 on AIX 5.2 ML3 I installed and setup Storage Agent on a W2K node for testing purposes. Everything tested out fine, and now I am wanting to remove everything, including the other server entry on the TSM server. I have already uninstalled the Storage Agent and device drivers from the W2K node. I have deleted all the filespaces related to that node name (a special node name setup to test the storage agent, the original node and filespaces still exist, but it never used the storage agent to do backups). I did all that last week sometime. Today, when I try to delete the other server entry for the Storage Agent, it tells me the server is currently in use. I deleted the nodename that we used for testing, and ran audit lic... no change. Nothing is listed in q ses related to this node. I checked netstat to verify there were no connections to that IP address... What am I missing? (I don't know if it matters, but we were testing LVSA and Image backups using the storage agent). TIA Todd
Re: Why can't I delete an other server entry?
Interesting... I verified that the define server command used serverpassword= instead of password=. This other server name was defined via the GUI anyway, so it would make sense that it used the correct syntax (one can assume). 02/11/05 14:31:28 ANR2017I Administrator MYID issued command: DEFINE SE RVER testserver COMMMETHOD=TCPIP SERVERPASSWORD=?***? HLA DDRESS=x.x.x.x LLADDRESS=1500 VALIDATEPROTOCOL=NO (S ESSION: 47407) Cycling the server process allowed me to delete the server. Thanks, Andy, Todd -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Andrew Raibeck Sent: Monday, February 21, 2005 2:24 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Why can't I delete an other server entry? OK, well, I just like to make sure all the bases are covered if they aren't explicitly mentioned... :-) APAR IC41610 might also be a match. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-02-21 12:56:31: Sorry.. Forgot to mention what I had checked out as a result of the ANR1664W message... I have no devtype definitions of server (at least, in the GUI, when I go to Objects, Server Storage, Device Classes, Server Device Classes, there are no entries)... We don't have an event server... And I addressed the connection issues with comments re: sessions and netstat connections.. Still can't figure it out.. Todd -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Andrew Raibeck Sent: Monday, February 21, 2005 1:26 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Why can't I delete an other server entry? Check the DELETE SERVER syntax, as it includes a list of items that will cause the command to fail. In particular, make sure you don't have any device SERVER-type classes that point to your other server. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-02-21 12:20:49: TSM 5.2.2.5 on AIX 5.2 ML3 I installed and setup Storage Agent on a W2K node for testing purposes. Everything tested out fine, and now I am wanting to remove everything, including the other server entry on the TSM server. I have already uninstalled the Storage Agent and device drivers from the W2K node. I have deleted all the filespaces related to that node name (a special node name setup to test the storage agent, the original node and filespaces still exist, but it never used the storage agent to do backups). I did all that last week sometime. Today, when I try to delete the other server entry for the Storage Agent, it tells me the server is currently in use. I deleted the nodename that we used for testing, and ran audit lic... no change. Nothing is listed in q ses related to this node. I checked netstat to verify there were no connections to that IP address... What am I missing? (I don't know if it matters, but we were testing LVSA and Image backups using the storage agent). TIA Todd
Re: Select for Tape Storage Pool Report
Here's something funny tho... Sometimes, when I run the script, I get a duplicate (sort of) entry in the output... Stg PoolTB PctUtl Logi Run? Recl Max VolUsed Diff -- --- -- -- - - --- - ..snipped.. L3_PRI_LTO6.70 57.4 99.6 NO 10035 29 6 L3_PRI_LTO6.70 57.4 99.6 NO 10035 134 ...chopped.. The run before this one showed the VolUsed column at 17 and 13 (still totaling 30). When I do a count of the volumes in that stgpool, it does come out to 30. The run after this one showed only a single entry for L3_PRI_LTO, with 30 tapes used. Odd behavior.. If anyone has a clue why, it would sure help satisfy my curiosity... Thx Todd -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Sung Y Lee Sent: Monday, February 21, 2005 9:47 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Select for Tape Storage Pool Report Very nice Even added the difference in counts. This select is very nice indeed. I think you over estimated this group about 93 seconds thing.. I tried to combine two selects for over an hour and gave up. Sung Y. Lee ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 02/21/2005 10:30:39 AM: Thanks, Here is the script in it's final form if any one else wants it. Adjust your devclass and char/decimal output and column titles as needed.. select - cast(a.stgpool_name as char(22)) as Stg Pool , - cast((a.est_capacity_mb/1024/1024) as dec(5,2)) as TB, - a.pct_utilized as PctUtl, - a.pct_logical as Logi, - cast(a.recl_running as char(4)) as Run?, - cast(a.reclaim as dec(3)) as Recl, - cast(a.maxscratch as dec(3)) as Max, - cast(count(*) as dec(3)) as VolUsed, - cast(a.maxscratch - count(*) as dec(3)) as Diff - from - stgpools a, - volumes b - where - devclass in ('LTO','LTO2') and - a.stgpool_name=b.stgpool_name - group by - a.stgpool_name, - a.est_capacity_mb, - a.pct_utilized, - a.pct_logical, - a.recl_running, - a.reclaim, - a.maxscratch -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Robert Ouzen Sent: Friday, February 18, 2005 11:44 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Select for Tape Storage Pool Report Hi Todd This Script will give you the number of volumes per stg and maxscratch allocated select a.stgpool_name,a.maxscratch,count(*) as Number of Vols from stgpools a, volumes b where a.stgpool_name = b.stgpool_name and a.devclass = 'SCALARCLASS' group by a.stgpool_name,a.maxscratch Regards Robert Ouzen Haifa University Israel -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Todd Lundstedt Sent: Saturday, February 19, 2005 12:00 AM To: ADSM-L@VM.MARIST.EDU Subject: Select for Tape Storage Pool Report I created this SQL select statement... select - cast(stgpool_name as char(22)) as Stg Pool , - cast((est_capacity_mb/1024/1024) as decimal(5,2)) as TB, - pct_utilized as Util, - pct_logical as Logi, - cast(recl_running as char(4)) as Run?, - cast(reclaim as dec(3)) as Recl, - cast(maxscratch as dec(3)) as Max - from stgpools - where devclass in ('LTO','LTO2') It outputs exactly what I think it should... Stg Pool TB Util LogiRun? Recl Max --------- -- L1_CPY_DBVL_LTO2_OFF1 19.06 14.5 100.0NO 100 55 L1_CPY_DB_LTO_OFF1 18.18 0.8 100.0NO 100 100 L1_CPY_DB_LTO_ON 0.00 0.0 100.0NO 100 25 L1_CPY_DSKIMG_LTO2_OFF 0.00 0.0 100.0NO 100 30 L1_CPY_LTO2_OFF110.91 3.9 99.9NO 100 30 L1_CPY_LTO_OFF1 9.50 35.9 99.5YES 60 50 L1_CPY_LTO_ON0.00 0.0 100.0NO 100 20 L1_PRI_DBVL_LTO216.65 16.6 100.0NO 100 40 L1_PRI_DB_LTO5.43 2.7 100.0NO 100 24 L1_PRI_DSKIMG_LTO2 0.00 0.0 100.0NO 60 10 L1_PRI_LTO 5.85 58.5 99.1NO 100 29 L1_PRI_LTO2 3.00 14.4 99.9NO 1008 L2_CPY_DB_LTO_OFF1 0.00 0.0 100.0NO 100 20 L2_CPY_DSKIMG_LTO2_OFF 0.00 0.0 100.0NO 100 30 L2_CPY_LTO_OFF1 5.07 22.2 99.8NO 60 30 L2_PRI_DB_LTO0.00 0.0 100.0NO 1001 L2_PRI_DSKIMG_LTO2 0.00 0.0 100.0NO 60 10 L2_PRI_LTO 2.32 48.8 99.6NO 100 15 L3_CPY_DB_LTO_OFF1 3.69 15.7 100.0YES 60
Select for Tape Storage Pool Report
I created this SQL select statement... select - cast(stgpool_name as char(22)) as Stg Pool , - cast((est_capacity_mb/1024/1024) as decimal(5,2)) as TB, - pct_utilized as Util, - pct_logical as Logi, - cast(recl_running as char(4)) as Run?, - cast(reclaim as dec(3)) as Recl, - cast(maxscratch as dec(3)) as Max - from stgpools - where devclass in ('LTO','LTO2') It outputs exactly what I think it should... Stg Pool TB Util LogiRun? Recl Max ----------- L1_CPY_DBVL_LTO2_OFF1 19.06 14.5 100.0NO100 55 L1_CPY_DB_LTO_OFF1 18.18 0.8 100.0NO100 100 L1_CPY_DB_LTO_ON 0.00 0.0 100.0NO100 25 L1_CPY_DSKIMG_LTO2_OFF 0.00 0.0 100.0NO100 30 L1_CPY_LTO2_OFF110.91 3.9 99.9NO100 30 L1_CPY_LTO_OFF1 9.50 35.9 99.5YES60 50 L1_CPY_LTO_ON0.00 0.0 100.0NO100 20 L1_PRI_DBVL_LTO216.65 16.6 100.0NO100 40 L1_PRI_DB_LTO5.43 2.7 100.0NO100 24 L1_PRI_DSKIMG_LTO2 0.00 0.0 100.0NO 60 10 L1_PRI_LTO 5.85 58.5 99.1NO100 29 L1_PRI_LTO2 3.00 14.4 99.9NO1008 L2_CPY_DB_LTO_OFF1 0.00 0.0 100.0NO100 20 L2_CPY_DSKIMG_LTO2_OFF 0.00 0.0 100.0NO100 30 L2_CPY_LTO_OFF1 5.07 22.2 99.8NO 60 30 L2_PRI_DB_LTO0.00 0.0 100.0NO1001 L2_PRI_DSKIMG_LTO2 0.00 0.0 100.0NO 60 10 L2_PRI_LTO 2.32 48.8 99.6NO100 15 L3_CPY_DB_LTO_OFF1 3.69 15.7 100.0YES60 20 L3_CPY_DOM_LTO_OFF1 4.54 11.1 100.0YES60 25 L3_CPY_DSKIMG_LTO2_OFF 0.00 0.0 100.0NO100 30 L3_CPY_LTO_OFF1 6.25 61.0 99.8NO 60 35 L3_CPY_MAIL_LTO_OFF1 3.56 5.7 100.0NO100 20 L3_PRI_DB_LTO2.64 22.1 100.0NO100 10 L3_PRI_DOM_LTO 1.74 29.0 100.0NO100 10 L3_PRI_DSKIMG_LTO2 0.00 0.0 100.0NO 60 10 L3_PRI_LTO 5.68 67.3 99.6NO100 30 L3_PRI_MAIL_LTO 1.77 17.0 100.0NO100 10 LTOTAPE 11.54 25.7 99.7NO 60 50 LTOTAPE_SQL 1.81 3.6 100.0NO100 10 OFFSITE 19.16 15.8 99.9YES60 100 SPACEMGPOOL 0.00 0.0 100.0NO 600 But, I want to add yet one more column.. the count of volumes currently assigned to that storage pool. I can find that information with the following select statement... select - count(*) - from volumes - where stgpool_name='SOME_STGPOOL_NAME' or.. select - stgpool_name, - count(*) - from volumes - group by stgpool_name I know I can select from multiple tables select a.col5, a.col20, b.col7 from table1 a, table2 b, blah blah.. But I am unsure how to code the count portion and relate it to the stgpool_name of that line... I also can't seem to get subselects to work.. I thought I had done this before, but I am not sure on that one... select * from (select col1, col2, col3 from tableA) I bet one of YOU can do it in 93 seconds, or less... The clock is ticking.. =) TIA Todd
TSM Windows Client 5.2.4?
Is there a release date on this? I am using 5.2.2.0 on a W2K node (server is 5.2.2.5 on AIX 5.2-03). Tried to do an Image backup and we got a few error messages during the backup. A quick search at ibm.com returned a single entry that matched my error almost to the letter, and indicated the problem has been fixed in the software version 5.2.4, but I don't see that version on the FTP site yet. It also indicated that a work around was to set resourceutilization to 1, and that seems to work, for now.
Re: DB2 LANFree reconfiguration to use LAN
We just moved some of our LANFree W2K and W2K3 clients from one datacenter to another. The SAN is still in the old datacenter, to LANFree from those clients is currently impossible (without additional hardware). I instructed the move team to stop, and disable the StorageAgent service on those machines that are moving, so the APIs use the default LAN path without changing any .opt files. Once the SAN is moved to the new datacenter, all we have to do is start the StorageAgent services, and life will (should) be good, again. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Daniel Sparrman Sent: Monday, January 24, 2005 9:02 AM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] DB2 LANFree reconfiguration to use LAN Depends on when the storage agent fails. If it fails during backup/restore processing, the DB/2 backup(or restore) will fail. If it fails and then the DB/2 process starts, yes it will fail-over to LAN functionality. Best Regards Daniel Sparrman --- Daniel Sparrman Exist i Stockholm 6B 183 62 TÄBY Växel: 08 - 754 98 00 Mobil: 070 - 399 27 51 Thomas, Matthew [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU 2005-01-24 15:57 Please respond to ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU To ADSM-L@VM.MARIST.EDU cc Subject Re: DB2 LANFree reconfiguration to use LAN Daniel (and Jeroen) - thanks for the advice. One further thought: If, for example, the StorageAgent failed or was disabled in Services, wouldn't the API default to the LAN path on finding the LANFree route unavailable? -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] Behalf Of Daniel Sparrman Sent: 24 January 2005 14:39 To: ADSM-L@VM.MARIST.EDU Subject: Re: DB2 LANFree reconfiguration to use LAN Yes, you will need to restart DB2 as it only reads the configuration file during startup Best Regards Daniel Sparrman --- Daniel Sparrman Exist i Stockholm AB Propellervägen 6B 183 62 TÄBY Växel: 08 - 754 98 00 Mobil: 070 - 399 27 51 Thomas, Matthew [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU 2005-01-24 15:36 Please respond to ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU To ADSM-L@VM.MARIST.EDU cc Subject DB2 LANFree reconfiguration to use LAN Guys, We're backing up DB2 v7 (WinNT) via LANFree to a TSM 4.2 server on AIX. All has been relatively fine and dandy. However, we're now due to upgrade the TSM server to v5.2 and must therefore revert to LAN backups for this DB2 instance. Does anyone know whether a db2stop/db2start is required to pick up the change from ENABLELANFREE YES to ENABLELANFREE NO in the api dsm.opt? TIA Matt Thomas Open Systems - Backup Midrange Storage Infrastructure Services --- This e-mail is intended only for the above addressee. It may contain privileged information. If you are not the addressee you must not copy, distribute, disclose or use any of the information in it. If you have received it in error please delete it and immediately notify the sender. evolvebank.com is a division of Lloyds TSB Bank plc. Lloyds TSB Bank plc, 25 Gresham Street, London, EC2V 7HN. Registered in England, number 2065. Telephone No: 020 7626 1500 Lloyds TSB Scotland plc, Henry Duncan House, 120 George Street, Edinburgh EH2 4LH. Registered in Scotland, number 95237. Telephone No: 0131 225 4555 Lloyds TSB Bank plc and Lloyds TSB Scotland plc are authorised and regulated by the Financial Services Authority and represent only the Scottish Widows and Lloyds TSB Marketing Group for life assurance, pensions and investment business. Signatories to the Banking Codes. --- --- This e-mail is intended only for the above addressee. It may contain privileged information. If you are not the addressee you must not copy, distribute, disclose or use any of the information in it. If you have received it in error please delete it and immediately notify the sender. evolvebank.com is a division of Lloyds TSB Bank plc. Lloyds TSB Bank plc, 25 Gresham Street, London, EC2V 7HN. Registered in England, number 2065. Telephone No: 020 7626 1500 Lloyds TSB Scotland plc, Henry Duncan House, 120 George Street, Edinburgh EH2 4LH. Registered in Scotland, number 95237. Telephone No: 0131 225 4555 Lloyds TSB Bank plc and Lloyds TSB Scotland plc are authorised and regulated by the Financial Services Authority and represent only the Scottish Widows and Lloyds TSB Marketing Group for life assurance, pensions and investment business. Signatories to the Banking Codes.
Re: resourceutilization
I thought resourceutilization was a Windows Client option only.. Did that change? -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Joni Moyer Sent: Monday, January 24, 2005 11:59 AM To: ADSM-L@VM.MARIST.EDU Subject: [ADSM-L] resourceutilization Hello All! I have a Solaris client at the 5.2.2.0 TSM client level and the user would like to know how to stream a single archive that sends multiple files to the TSM server through 1 session instead of multiple sessions. I have the MAXNUMMP set to 2 for this client and I have looked at the dsm.sys file and there is no resourceutilization parameter within the file, so I believe it is using the default. I couldn't find the value for the default of the number of sessions a client can start on the TSM server, which in my case is an AIX 5.2.2.5 server. Here is the user's question: If I issue an archive with multiple files TSM starts multiple sessions - one for each file. Do we have a approach to limit the number of sessions? Any suggestions/help would be appreciated. Also, when issuing an archive, is it possible to override the dsm.sys file and specify the resourceutilization for that archive at the time it is issue? Thanks again! Joni Moyer Highmark Storage Systems Work:(717)302-6603 Fax:(717)302-5974 [EMAIL PROTECTED]
Re: resourceutilization
Found in the Windows version of Backup Archive Clients Installation and User's Guide.pdf, v5.2, page 329, RESOURceutilizaion value... Supported Clients This option is valid for all Windows clients. The server can also define this option. The Tivoli Storage Manager client API does not support this option. I read that as it is only valid in Windows clients, especially since the base API doesn't support the value... Do I interpret this incorrectly? I also find no mention of the option in any of the AIX client pdfs that I have around. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Andrew Raibeck Sent: Monday, January 24, 2005 12:31 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] resourceutilization Nothing has changed. When this option was introduced, it was (and still is) available for all clients. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-01-24 11:13:46: I thought resourceutilization was a Windows Client option only.. Did that change? -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Joni Moyer Sent: Monday, January 24, 2005 11:59 AM To: ADSM-L@VM.MARIST.EDU Subject: [ADSM-L] resourceutilization Hello All! I have a Solaris client at the 5.2.2.0 TSM client level and the user would like to know how to stream a single archive that sends multiple files to the TSM server through 1 session instead of multiple sessions. I have the MAXNUMMP set to 2 for this client and I have looked at the dsm.sys file and there is no resourceutilization parameter within the file, so I believe it is using the default. I couldn't find the value for the default of the number of sessions a client can start on the TSM server, which in my case is an AIX 5.2.2.5 server. Here is the user's question: If I issue an archive with multiple files TSM starts multiple sessions - one for each file. Do we have a approach to limit the number of sessions? Any suggestions/help would be appreciated. Also, when issuing an archive, is it possible to override the dsm.sys file and specify the resourceutilization for that archive at the time it is issue? Thanks again! Joni Moyer Highmark Storage Systems Work:(717)302-6603 Fax:(717)302-5974 [EMAIL PROTECTED]
Re: resourceutilization
OK.. I will look at the manuals again, and see what I have downloaded.. Must have mislabeled one, or something. Thanks -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Andrew Raibeck Sent: Monday, January 24, 2005 1:03 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] resourceutilization I can see how that might be misleading, but the Windows client manual information is in the context of the Windows OS, and not necessarily that of other operating systems (Unix, NetWare, etc.). If you look in the corresponding manual for the other OSes, you'll see that RESOURCEUTILIZATION appears there, too. I don't know which PDF files you are looking at, but I verified that it goes back at least as far as the version 4.2 books (oldest manual version I have readily available) for Mac, Unix, and NetWare. RESOURCEUTILIZATION is not available in the API because that option is used to funnel data at the file level, and the TSM API does not perform any file I/O. It would be up to the vendor of an API application to implement their own RESOURCEUTILIZATION-like functionality in their application. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-01-24 11:46:56: Found in the Windows version of Backup Archive Clients Installation and User's Guide.pdf, v5.2, page 329, RESOURceutilizaion value... Supported Clients This option is valid for all Windows clients. The server can also define this option. The Tivoli Storage Manager client API does not support this option. I read that as it is only valid in Windows clients, especially since the base API doesn't support the value... Do I interpret this incorrectly? I also find no mention of the option in any of the AIX client pdfs that I have around. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Andrew Raibeck Sent: Monday, January 24, 2005 12:31 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] resourceutilization Nothing has changed. When this option was introduced, it was (and still is) available for all clients. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-01-24 11:13:46: I thought resourceutilization was a Windows Client option only.. Did that change? -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Joni Moyer Sent: Monday, January 24, 2005 11:59 AM To: ADSM-L@VM.MARIST.EDU Subject: [ADSM-L] resourceutilization Hello All! I have a Solaris client at the 5.2.2.0 TSM client level and the user would like to know how to stream a single archive that sends multiple files to the TSM server through 1 session instead of multiple sessions. I have the MAXNUMMP set to 2 for this client and I have looked at the dsm.sys file and there is no resourceutilization parameter within the file, so I believe it is using the default. I couldn't find the value for the default of the number of sessions a client can start on the TSM server, which in my case is an AIX 5.2.2.5 server. Here is the user's question: If I issue an archive with multiple files TSM starts multiple sessions - one for each file. Do we have a approach to limit the number of sessions? Any suggestions/help would be appreciated. Also, when issuing an archive, is it possible to override the dsm.sys file and specify the resourceutilization for that archive at the time it is issue? Thanks again! Joni Moyer Highmark Storage Systems Work:(717)302-6603 Fax:(717)302-5974 [EMAIL PROTECTED]
Re: Filespace query
Shouldn't you be using the query occupancy command? -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Jones, Eric J Sent: Monday, January 24, 2005 12:54 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: [ADSM-L] Filespace query Thanks. I can now query the various drives but I cannot seem to get the SYSTEM OBJECT. I want to be able to tell how much space the SYSTEM OBJECT is taking up(90 days worth). With the space in the name it does not seem to find it. I tried using single quotes and double quotes around SYSTEM OBJECT, single and double quotes around the entire filespace(\\owgttsm\SYSTEM OBJECT) and it does not seem to find it. Thanks again, Eric -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] Behalf Of Andrew Raibeck Sent: Monday, January 24, 2005 1:14 PM To: ADSM-L@VM.MARIST.EDU Subject: Re: Filespace query Try tacking on the NAMETYPE=UNICODE option on the QUERY FILESPACE command. Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. ADSM: Dist Stor Manager ADSM-L@VM.MARIST.EDU wrote on 2005-01-24 10:52:44: Good Afternoon. Is there an easy way to see how much space the SYSTEM OBJECT is taking up on a particular client? I can see what the total space that the client uses but I'd like to find out the individual filespaces(total). I tried to do aquery filespace owgwttsm SYSTEM OBJECT f=d and I get File space note found. I then did a query on the d drive query filespace owgwttsm \\owgttsm\d$ f=d (matches what I see for the filespace under the client name and I receive Filespace not found. == I'm running AIX 5.2 TSM5.2.2 on the server and the client is win2K SP3 with TSM 5.2.2 Have a Great Day, Eric
[no subject]
We will be moving our 3584 library (including the TSM server, shark, and everything else in our data center) across town to a new facility. I am trying to decide what the best method is to accomplish certain aspects of the library move. We don't want to leave the tapes in the library during the move. Is it best to checkout the tapes before the move, or just remove the tapes, and put them back into the library after the move in no specific order, and run an audit library libname checklabel=barcode after the move? If we go the checkout route, I have a script to remove the scratch tapes first using the I/O slots. We will have to run the script several times (10 slots), until all the scratch tapes are removed. Then, I planned on creating and executing a script that for each VOLNAME in a query libv command, it would do a checkout libv libname $VOLNAME remove=no checklabel=no. The way I understand the help commands, this will leave all the existing volumes (which should be only in use volumes) in their current SCSI slot. We would then remove and box up all the tapes. At the new datacenter, we would place all the tapes in no particular order directly back into the library storage slots, then I would run a checkin libv libname search=yes checklabel=barcode status=private. After that completed, I would run through several checkin libv commands to put the scratch volumes back into the library through the I/O slots. All that sounds like a lot of work when, in theory, I should be able to just shut down the TSM server, remove the tapes (scratch and private together), transport the tapes and library (and everything), re-insert the tapes, start TSM and run an audit library libname checklabel=barcode as soon as it comes up... right? Of course, I would, via the 3584 User Panel or the 3584 Web Interface, remove the cleaning cartridges and diagnosis cartridge after TSM shuts down, and insert them before TSM starts up again. Thoughts? Suggestions? Horror stories? ps, please reply directly to me, as well as the listserv. I never did unsubscribe to the listserv, but I have not been getting any email from the listserv in several months. I subscribed again today, so we will see how it goes. Thanks in advance, Todd
Multiple nodes in dsm.opt (or dsm.sys).
Client is 5.2.2.9 running on AIX5.2 with HACMP. On AIX, how do I setup the baclient to recognize more than one node name, so I can start multiple dsmc sched processes; one for the local drives, and one for each shared resource group (I have made this work on Windows boxes, but this is the first time I have had to do this on AIX). I had intended to keep the dsm.opt, dsm.sys, schedule logs, inclexcl list, and error logs for the local filesystems in /usr/tivoli/tsm, and the dsm.opt, dsm.sys, logs and list for the shared resource groups on the shared disks. I need to be able to have different inclexcl lists for each nodename to ensure things get backed up to the correct nodename. ++The following two files (plus an inclexcl.lst file) exist in /usr/tivoli/tsm/client/ba/bin --dsm.opt for local hard drive files: SErvername axhrtlp1 --dsm.sys for local hard drive files: SErvername axhrtlp1 COMMmethod TCPip TCPPort1500 TCPServeraddress 200.1.1.41 PASSWORDAccess generate schedlogname /usr/tivoli/tsm/client/ba/bin/dsmsched.log INCLEXCL /usr/tivoli/tsm/client/ba/bin/inclexcl.lst errorlogname /usr/tivoli/tsm/client/ba/bin/dsmerror.log schedlogretention 5 errorlogretention 5 DIRMC VCHS_DIRMC ++The following two files (plus an inclexcl.lst file) exist in /oracle/cluster/tivoli/tsm/client/ba/bin --dsm.opt for shared drive files: SErvername axhrorap --dsm.sys for shared drive files: SErvername axhrorap COMMmethod TCPip TCPPort1500 TCPServeraddress 200.1.1.41 PASSWORDAccess generate schedlogname /oracle/cluster/tivoli/tsm/client/ba/bin/dsmsched.log INCLEXCL /oracle/cluster/tivoli/tsm/client/ba/bin/inclexcl.lst errorlogname /oracle/cluster/tivoli/tsm/client/ba/bin/dsmerror.log schedlogretention 5 errorlogretention 5 DIRMC VCHS_DIRMC When I execute dsmc -optfile=/oracle/cluster/tivoli/tsm/client/ba/bin/dsm.opt and enter the servername of axhrorap, I get the following error... 08/12/04 15:28:18 ANS1370E The user ID entered does not have the correct level of authority to access the client machine. I searched through the documentation, and nothing jumped out at me as a solution. Where can I read about how to set this up? Thanks in advance, Todd
Multiple copy pools per primary pool
I am just trying to consider all the pros/cons of having multiple copy storage pools per primary pool. I was considering having one copy storage pool that never gets taken offsite (remains in the library), and two copy pools that get checked out and set to offsite. Obviously, the pros are that, in the event of a disaster, you have multiple offsite storage pools offsite, and volumes that are destroyed, damaged, or lost in transit to your recovery center are duplicated in the storage pool set that remained offsite. Additionally, the onsite copy storage pool can be used to restore damaged primary storage pool tapes. For a recent disaster recovery test, I made a second copy of an offsite storage pool, so we wouldn't have to take the only offsite tapes that existed. During the recovery process (for which I was creating the procedures on the fly), I ran into a snag. The secondary copy pool tapes I took didn't have all of the data that existed in its primary, or the real offsite copy pool (this was known ahead of time). I set the volumes in the primary storage pool to destroyed, and tried to recover some files. When I got to the point where the files the client wished to recover were only in the copy pool that was back home, TSM requested we checkin, and mount that volume. At the time, my work around was to delete the volumes (discarddata=yes) in the storage pool that was back home, since an access of destroyed is not allowed for copy pool volumes. This allowed me to restore the files. I am now wondering if simply setting the volumes to unavailable instead of offsite would have worked. If you have input on this, I would like to hear it. According to the TSM Help system, if you set a sequential volume to unavailable, TSM will not attempt to mount the volume. That sounds pretty cut and dry, but without a safe way to test that, I don't want to include that in any procedures I create for DR. Thanks in advance, Todd
TDP, Exchange 2003, and SANs
I am working up my plan to backup our new Exchange 2003 storage groups. Our environment includes: Exchange 2003 running on Windows 2003 TSM 5.2.2.5 running on AIX 5.2 Shark storage for the AIX server's disk based storage EVA (I think) storage for the Exchange server data LTO1 in a 3584 accessed across a McData 4500 SAN switch I read through the TDP for Exchange v5.2 documentation. It appears that I cannot split up the different data types (meta, data, logs) into different management classes, but I can put the different backup types (full, copy, incr, diff, dbcopy) into different management classes. My questions are... When doing an incr, or diff backup via TDP, how does TDP do that? I intend to backup the storage groups full across the SAN. The storage groups will be 20GB in size. From what the Exchange admins are telling me, Exchange creates log files with a max size of 5MB. During an incr or diff backup, does TDP backup all the log file transaction information as one filename to TSM, allowing for less overhead when using a SAN, or does TDP keep individual log file names intact, backing up each one serially (which would lead me to believe that using the SAN would actually slow things down)? All my experience with TDP is in SQL (and a bit in Domino). Does TDP for Exchange get out of sync as easily as TDP for SQL when dealing with management class name changes? With TDP for SQL, I have to rename the existing filespaces and start with new filespaces whenever I change management class names to ensure the meta stays in sync with the data. Are there any quirks like this in TDP for Exchange? Are there any other things I should be considering? Thanks in advance, Todd
Re: TDP, Exchange 2003, and SANs
Yeah.. I read that in the documentation, and understand that log files are deleted after a successful incremental, and not deleted after a successful diff... But, I am looking for more detail in the actual transfer and storage, as the rest of my email indicated. Thx though Stapleton, Mark [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 06/24/2004 09:45 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: TDP, Exchange 2003, and SANs From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] On Behalf Of Todd Lundstedt I read through the TDP for Exchange v5.2 documentation. It appears that I cannot split up the different data types (meta, data, logs) into different management classes, but I can put the different backup types (full, copy, incr, diff, dbcopy) into different management classes. When doing an incr, or diff backup via TDP, how does TDP do that? An incremental backup with TDP for Mail (Exchange) is actually a backup of all log files since the last incremental backup. A differential backup is, IIRC, a backup of the all the log files since the last full backup. -- Mark Stapleton
Re: TDP, Exchange 2003, and SANs
Del, Thanks.. follow up question... You said... Data Protection for Exchange performs incremental and differential backups by backing up the full transaction log files to TSM. They are all placed into a single TSM backup object. From this, would it be safe to assume that the data from hundreds of 5MB log files would stream well across the SAN, since your description infers that the incr would be handled with a single TSM database transaction? Does the txngroupmax value have any bearing on this? You said the incr/diff backup only backs up the full transaction log files. That concerns me a bit. The Exchange admin indicated that Exchange has four log files open at all times. This means that up to 20 MB of transactions would not get backed up during a backup (full included). We are new to Exchange, migrating from Domino (plea$e, plea$e, don't a$k me why we would even what to do that... it'$ a long $tory). I guess we will have to just live with the incomplete/incorrect point of recovery. Thanks again for the quick and informative reply, Todd Del Hoobler [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 06/24/2004 10:02 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: TDP, Exchange 2003, and SANs Todd, I have embedded some answers below. Thanks, Del ADSM: Dist Stor Manager [EMAIL PROTECTED] wrote on 06/24/2004 10:39:59 AM: I am working up my plan to backup our new Exchange 2003 storage groups. Our environment includes: Exchange 2003 running on Windows 2003 TSM 5.2.2.5 running on AIX 5.2 Shark storage for the AIX server's disk based storage EVA (I think) storage for the Exchange server data LTO1 in a 3584 accessed across a McData 4500 SAN switch I read through the TDP for Exchange v5.2 documentation. It appears that I cannot split up the different data types (meta, data, logs) into different management classes, but I can put the different backup types (full, copy, incr, diff, dbcopy) into different management classes. This is intentional because you want all TSM objects that represent a single, logical backup to be managed under the same policy but you may also want to be able to manage the different backup types with different policy. My questions are... When doing an incr, or diff backup via TDP, how does TDP do that? I intend to backup the storage groups full across the SAN. The storage groups will be 20GB in size. From what the Exchange admins are telling me, Exchange creates log files with a max size of 5MB. During an incr or diff backup, does TDP backup all the log file transaction information as one filename to TSM, allowing for less overhead when using a SAN, or does TDP keep individual log file names intact, backing up each one serially (which would lead me to believe that using the SAN would actually slow things down)? Data Protection for Exchange performs incremental and differential backups by backing up the full transaction log files to TSM. They are all placed into a single TSM backup object. During restore, Data Protection for Exchange will pull the individual log files out of the single TSM object and write them to disk. All my experience with TDP is in SQL (and a bit in Domino). Does TDP for Exchange get out of sync as easily as TDP for SQL when dealing with management class name changes? With TDP for SQL, I have to rename the existing filespaces and start with new filespaces whenever I change management class names to ensure the meta stays in sync with the data. Are there any quirks like this in TDP for Exchange? SQL has some unique issues that cause the anomaly that you are referring to. The root cause of it on SQL is the multiple-striping that is allowed and the temporary objects that are used to hold pre-transaction-commit striped data. Even if you didn't use multiple-striping, it can occur. Data Protection for Exchange does NOT have this issue. Are there any other things I should be considering? Thanks in advance, Todd
Re: TDP, Exchange 2003, and SANs
Thanks, Del. My comment regarding living with an incomplete/incorrect point of recovery was solely meant to represent that in a complete data center wipe-out, the existing full and incrementals that we have offsite would not accurately represent what was in a particular mailbox at the time of the backup. Since TDP doesn't backup a log file unless it is filled (and I assume that means if a log file is open, the TDP doesn't backup that information in those logs). If somehow, TDP is able to get this information from the currently open log files (four per storage group, or is it four per mail store), then we will be OK. Thanks again.. I will be testing how all this works as soon as we can. Todd Del Hoobler [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 06/24/2004 10:39 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: TDP, Exchange 2003, and SANs Todd, Answers embedded. Thanks, Del ADSM: Dist Stor Manager [EMAIL PROTECTED] wrote on 06/24/2004 11:17:55 AM: Del, Thanks.. follow up question... You said... Data Protection for Exchange performs incremental and differential backups by backing up the full transaction log files to TSM. They are all placed into a single TSM backup object. From this, would it be safe to assume that the data from hundreds of 5MB log files would stream well across the SAN, since your description infers that the incr would be handled with a single TSM database transaction? Does the txngroupmax value have any bearing on this? There are always various factors that can influence this, however, in theory, you are correct. These will be represented by a single stream of data being backing up to a single TSM object. txngroupmax will not affect this since it is a single TSM object. My best suggestion is to try it to see how it performs for you. You said the incr/diff backup only backs up the full transaction log files. That concerns me a bit. The Exchange admin indicated that Exchange has four log files open at all times. This means that up to 20 MB of transactions would not get backed up during a backup (full included). We are new to Exchange, migrating from Domino (plea$e, plea$e, don't a$k me why we would even what to do that... it'$ a long $tory). I guess we will have to just live with the incomplete/incorrect point of recovery. You don't need to live with anything. You don't need to be concerned, either. Data Protection for Exchange works directly in conjunction with the Exchange server through the Exchange backup API. The Exchange Server tells Data Protection for Exchange which logs files need to be backed up. Data Protection for Exchange takes that list and backs them up. You don't need to worry about an incomplete backup, either. Since Data Protection for Exchange uses the Exchange backup API, all of your backups will be completely consistent. Thanks again for the quick and informative reply, Todd
Move Nodedata starts and ends without doing anything
Recently, I moved data for several nodes from one Primary storage pool to another. That new primary storage pool is backing up to a different copy storage pool. Data for those servers in the original copy storage pool (stgpool_name='OFFSITE') is no longer needed. So, I was going to perform a move nodedata for those servers with FROMstgpool=OFFSITE (TOstgpool=OFFSITE is assumed since this is a copy storage pool) and then delete the new OFFSITE storage pool volumes with discarddata=yes to remove them from the old copy storage pool. But nothing moved. What is wrong? Thanks in advance, Todd + TSM Server 5.1.7.1 running on AIX 4.3.3. relevant storage pool info: Storage Pool Name: OFFSITE Storage Pool Type: Copy Device Class Name: LTO Reclamation Threshold: 100 Maximum Scratch Volumes Allowed: 100 Storage Pool Data Format: Native Total volumes used in stgpool currently: 71, all access=offsite That leaves 29 volumes available to be added to the storage pool. There are 19 scratch volumes in the library. Q OCC NT-DCO-EPRISE1 STG=OFFSITE (sorry if this ends up wrapping) Node Name Type Filespace Storage Number of PhysicalLogical NamePool Name Files Space Space Occupied Occupied (MB) (MB) -- -- -- - - - NT-DCO-EP- Bkup \\nt-dco-- OFFSITE39,645 3,990.87 3,950.56 RISE1 eprise1\- g$ NT-DCO-EP- Bkup \\nt-dco-- OFFSITE 9,567 493.96 475.78 RISE1 eprise1\- c$ 11/11/03 09:51:56 ANR2017I Administrator TLUNDSTE issued command: MOVE NODEDATA NT-DCO-EPRISE1 from=offsite maxprocess=1 reconstr=yes 11/11/03 09:51:56 ANR1643I MOVE NODEDATA: All file spaces for node NT-DCO-EPRISE1, will be moved. 11/11/03 09:51:56 ANR0984I Process 2579 for MOVE NODE DATA started in the BACKGROUND at 09:51:56. 11/11/03 09:51:56 ANR1284I Move node data started as process 2579. 11/11/03 09:51:56 ANR2110I MOVE NODEDATA started as process 2579. 11/11/03 09:51:56 ANR0609I MOVE NODEDATA started as process 2579. 11/11/03 09:51:56 ANR1288I Move node data process 2579 ended for storage pool OFFSITE. 11/11/03 09:51:56 ANR0985I Process 2579 for MOVE NODE DATA running in the BACKGROUND completed with completion state SUCCESS at 09:51:56. 11/11/03 09:51:56 ANR1290I Move node data from storage pool OFFSITE to storage pool OFFSITE has ended. Files Moved: 0, Bytes Moved: 0, Unreadable Files: 0, Unreadable Bytes: 0. +
Re: Move Nodedata starts and ends without doing anything
Shouldn't it use primary storage volumes to get the data, just like a reclamation or a move data for acc=offsite volumes? But nothing moved. What is wrong? The tapes are offsite and thus the data can't be moved?
Re: Move Nodedata starts and ends without doing anything
Well, there it is then. We'll see if TSM support comes back with this bit of information on their next attempt to resolve this issue. Their first attempt was to tell me that I couldn't move nodedata from a copy storage pool to another storage pool. I guess they didn't read the entire ESR. Anyway, I agree... a note about volume access being readwrite or readonly most certainly could be added to the help move nodedata text, as well. Thanks, Richard. ps. The Technical Guide redbook now has a direct link on my desktop. Thanks. Shouldn't it use primary storage volumes to get the data, just like a reclamation or a move data for acc=offsite volumes? Again, I urge reading the Technical Guide redbook to fully understand TSM functions before trying to use them. It's particularly important to do so because the command reference manual ends up getting just a fraction of the information presented in the TG - often omitting important facts, such as: The volumes added to this list have to have an access of READWRITE or READONLY. That obviously precludes Access=Offsite. And wouldn't it be nice if the programmers put out some kind of message when the function ends up copying none of the data that the customer requested? (Come on, IBM - that's not enterprise level programming.) Richard Sims, BU
Re: How to mark all TAPEPOOL volumes read only
Or.. (if your storage pool name is TAPEPOOL, then) update volume * access=readonly wherestgpool=tapepool goran [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 11/05/2003 09:16 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: How to mark all TAPEPOOL volumes read only update volume _volume_name acc=reado - Original Message - From: Farren Minns [EMAIL PROTECTED] To: [EMAIL PROTECTED] Sent: Wednesday, November 05, 2003 4:08 PM Subject: How to mark all TAPEPOOL volumes read only Hi TSMers TSM 5.1.6.2 on Solaris 2.7 What's the best way to make all tapes in our TAPEPOOL read-only. I don't want the TAPEPOOL itself read-only, just the volumes it's using at present. Then it can start using new scratch tapes from now on. Thanks Farren Minns - John Wiley Sons Ltd * This email transmission is confidential and intended for the person or organisation it is addressed to. If you are not the intended recipient, you must not copy, distribute, or disseminate the information, open any attachment, or take any action in reliance of it. If you have received this message in error please notify the sender. Any views expressed in this message are those of the individual sender, except where the sender specifically states otherwise. Although this email has been scanned for viruses you should rely on your own virus check, as the sender takes no responsibility for any damage arising out of any bug or virus infection. *
Tapes being set to Unavailable (was Re: 3584 Cleaning Frequency)
IBM.. and it is happening with frequency on tapes that are new, as well as tapes we have had since day one. I cleaned the drives manually (using the panel on the 3584), but we are still getting a few unavailables each night. I did some research on the tapes that went unavailable over the last 30 days, and all of the ones I checked had another error nearby.. indicating that the dismount of that volume failed, and it may still be in the drive... yet, they never are, they seem to be dismounted successfully. I am still trying to figure this out... if anyone has a suggestion. Thanks Lambelet,Rene,VEVEY,GL-CSC [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 10/16/2003 07:35 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: 3584 Cleaning Frequency Todd, what kind of tapes do you use (Imation, 3M, IBM, Fuji ? We had much much more r/o tapes with Imation than with IBM ones ! rené -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED] Sent: Monday,13. October 2003 18:36 To: [EMAIL PROTECTED] Subject: 3584 Cleaning Frequency What is the normal clean frequency for LTO-1 drives in a 3584? I have the drives defined to TSM with a cleaning frequency of none, and the 3584 library set to Automatic cleaning, per recommendations. Yet, when I use the front panel of the 3584 to look at the cleaning cartridges, one has never been used, and the other has only been used 4 times. We have had this library for nearly 2 years now. I started looking at this because of the number of tapes that are marked read-only or unavailable per night (anywhere from one to 7 so far, usually 2 or 3 per night). The eight drives show as follows: Drive1 Library Loads: 10391 Library Unloads:10381 Drive Loads:867 (drive1 has been reset recently) MB Written: 2903279 MB Read:1923642 Cleans: 0 -- Drive2 Library Loads: 10023 Library Unloads:10015 Drive Loads:6535 MB Written: 49313622 MB Read:26764547 Cleans: 0 -- Drive3 Library Loads: 10239 Library Unloads:10234 Drive Loads:1598 MB Written: 10083308 MB Read:5886656 Cleans: 0 -- Drive4 Library Loads: 6877 Library Unloads:6873 Drive Loads:6461 MB Written: 44438571 MB Read:25808491 Cleans: 0 -- Drive5 Library Loads: 6528 Library Unloads:6527 Drive Loads:6705 MB Written: 47722355 MB Read:28614867 Cleans: 0 -- Drive6 Library Loads: 1933 Library Unloads:1932 Drive Loads:2144 MB Written: 11874084 MB Read:6888474 Cleans: 0 -- Drive7 Library Loads: 2051 Library Unloads:2050 Drive Loads:2823 MB Written: 27765506 MB Read:7691746 Cleans: 0 -- Drive8 Library Loads: 2102 Library Unloads:2101 Drive Loads:2265 MB Written: 12299072 MB Read:8002016 Cleans: 0 These tape drives have transferred thousands of terabytes of data (what's that.. exabyte?), apparently without cleanings. What can I check to find out why, other than what I mentioned above (Lib set to AUTO, and drives set in TSM to none). Thanks in advance Todd
Re: Tapes being set to Unavailable (was Re: 3584 Cleaning Frequency)
Yep.. they are at a very recent, if not the most recent code level, along with the library's code level. No more than 2 months old there... but then, that could be the problem.. perhaps that code level is bad.. I will have my CE get me the latest numbers and check that, but I suspect (hope) that isn't the issue. David Longo [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 10/16/2003 02:11 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Tapes being set to Unavailable (was Re: 3584 Cleaning Frequency) Have you had the firmware on the drives brought uptodate? David B. Longo System Administrator Health First, Inc. 3300 Fiske Blvd. Rockledge, FL 32955-4305 PH 321.434.5536 Pager 321.634.8230 Fax:321.434.5509 [EMAIL PROTECTED] [EMAIL PROTECTED] 10/16/03 02:43PM IBM.. and it is happening with frequency on tapes that are new, as well as tapes we have had since day one. I cleaned the drives manually (using the panel on the 3584), but we are still getting a few unavailables each night. I did some research on the tapes that went unavailable over the last 30 days, and all of the ones I checked had another error nearby.. indicating that the dismount of that volume failed, and it may still be in the drive... yet, they never are, they seem to be dismounted successfully. I am still trying to figure this out... if anyone has a suggestion. Thanks Lambelet,Rene,VEVEY,GL-CSC [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 10/16/2003 07:35 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: 3584 Cleaning Frequency Todd, what kind of tapes do you use (Imation, 3M, IBM, Fuji ? We had much much more r/o tapes with Imation than with IBM ones ! rené -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED] Sent: Monday,13. October 2003 18:36 To: [EMAIL PROTECTED] Subject: 3584 Cleaning Frequency What is the normal clean frequency for LTO-1 drives in a 3584? I have the drives defined to TSM with a cleaning frequency of none, and the 3584 library set to Automatic cleaning, per recommendations. Yet, when I use the front panel of the 3584 to look at the cleaning cartridges, one has never been used, and the other has only been used 4 times. We have had this library for nearly 2 years now. I started looking at this because of the number of tapes that are marked read-only or unavailable per night (anywhere from one to 7 so far, usually 2 or 3 per night). The eight drives show as follows: Drive1 Library Loads: 10391 Library Unloads:10381 Drive Loads:867 (drive1 has been reset recently) MB Written: 2903279 MB Read:1923642 Cleans: 0 -- Drive2 Library Loads: 10023 Library Unloads:10015 Drive Loads:6535 MB Written: 49313622 MB Read:26764547 Cleans: 0 -- Drive3 Library Loads: 10239 Library Unloads:10234 Drive Loads:1598 MB Written: 10083308 MB Read:5886656 Cleans: 0 -- Drive4 Library Loads: 6877 Library Unloads:6873 Drive Loads:6461 MB Written: 44438571 MB Read:25808491 Cleans: 0 -- Drive5 Library Loads: 6528 Library Unloads:6527 Drive Loads:6705 MB Written: 47722355 MB Read:28614867 Cleans: 0 -- Drive6 Library Loads: 1933 Library Unloads:1932 Drive Loads:2144 MB Written: 11874084 MB Read:6888474 Cleans: 0 -- Drive7 Library Loads: 2051 Library Unloads:2050 Drive Loads:2823 MB Written: 27765506 MB Read:7691746 Cleans: 0 -- Drive8 Library Loads: 2102 Library Unloads:2101 Drive Loads:2265 MB Written: 12299072 MB Read:8002016 Cleans: 0 These tape drives have transferred thousands of terabytes of data (what's that.. exabyte?), apparently without cleanings. What can I check to find out why, other than what I mentioned above (Lib set to AUTO, and drives set in TSM to none). Thanks in advance Todd ## This message is for the named person's use only. It may contain confidential, proprietary, or legally privileged information. No confidentiality or privilege is waived or lost by any mistransmission. If you receive this message in error, please immediately delete it and all copies of it from your system, destroy any hard copies of it, and notify the sender. You must not, directly or indirectly, use, disclose, distribute, print
Re: Tapes being set to Unavailable (was Re: 3584 Cleaning Frequency)
Clips?? Please explain. John Monahan [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 10/16/2003 03:44 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: Tapes being set to Unavailable (was Re: 3584 Cleaning Frequency) Has your CE installed the new drive clips? That helps with the dismount problems. __ John Monahan Senior Consultant Enterprise Solutions Computech Resources, Inc. Office: 952-833-0930 ext 109 Cell: 952-221-6938 http://www.compures.com Todd Lundstedt [EMAIL PROTECTED] IA-CHRISTI.ORGTo Sent by: ADSM: [EMAIL PROTECTED] Dist Stor cc Manager [EMAIL PROTECTED] Subject .EDU Re: Tapes being set to Unavailable (was Re: 3584 Cleaning Frequency) 10/16/2003 02:16 PM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] .EDU Yep.. they are at a very recent, if not the most recent code level, along with the library's code level. No more than 2 months old there... but then, that could be the problem.. perhaps that code level is bad.. I will have my CE get me the latest numbers and check that, but I suspect (hope) that isn't the issue. David Longo [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 10/16/2003 02:11 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Tapes being set to Unavailable (was Re: 3584 Cleaning Frequency) Have you had the firmware on the drives brought uptodate? David B. Longo System Administrator Health First, Inc. 3300 Fiske Blvd. Rockledge, FL 32955-4305 PH 321.434.5536 Pager 321.634.8230 Fax:321.434.5509 [EMAIL PROTECTED] [EMAIL PROTECTED] 10/16/03 02:43PM IBM.. and it is happening with frequency on tapes that are new, as well as tapes we have had since day one. I cleaned the drives manually (using the panel on the 3584), but we are still getting a few unavailables each night. I did some research on the tapes that went unavailable over the last 30 days, and all of the ones I checked had another error nearby.. indicating that the dismount of that volume failed, and it may still be in the drive... yet, they never are, they seem to be dismounted successfully. I am still trying to figure this out... if anyone has a suggestion. Thanks
Re: I/O errors
TSM 5.1.7.1 on AIX 4.3.3 ML10, Fiber Channel HBA, IBM LTO1 We recently upgraded from 4.2.1.7 to 5.1.7.1. I started seeing the unavailable and read-only tapes soon after that. I haven't done anything else to determine the issues. I thought it might be due to dirty tape drives (recently sent an email to the list regarding that), and manually forced a clean of all the drives, but that didn't resolved the problem. Krzysztof Kus [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 10/14/2003 10:22 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:I/O errors Hello! My configuration is: p6H0 with AIX 4.3.3 ML 9 TSM 5.1.6.1 HBA 6228 bridge FC-SCSI 3250 drives HP-Ultrium1 From time to time randomly I get error messages in errpt ADSM_DD_LOG2. From the sense data I can see that the problem is with a medium. In the TSM actlog there are messages I/O error..., and the tape is marked as read-only (I don't include error logs because I want my mail to be read). I've checked almost everything from tapes, microcode on the drives through optical fibers to AIX and TSM (there is no errors on the bridge). Has anybody have similar adventure with TSM? I'm opened for every suggestions. Next week I'm going to connect my library and drives directly through the SCSI (this week I strongly suspect the HBA) :) Regards. -- --== KKus ==--
3584 Cleaning Frequency
What is the normal clean frequency for LTO-1 drives in a 3584? I have the drives defined to TSM with a cleaning frequency of none, and the 3584 library set to Automatic cleaning, per recommendations. Yet, when I use the front panel of the 3584 to look at the cleaning cartridges, one has never been used, and the other has only been used 4 times. We have had this library for nearly 2 years now. I started looking at this because of the number of tapes that are marked read-only or unavailable per night (anywhere from one to 7 so far, usually 2 or 3 per night). The eight drives show as follows: Drive1 Library Loads: 10391 Library Unloads:10381 Drive Loads:867 (drive1 has been reset recently) MB Written: 2903279 MB Read:1923642 Cleans: 0 -- Drive2 Library Loads: 10023 Library Unloads:10015 Drive Loads:6535 MB Written: 49313622 MB Read:26764547 Cleans: 0 -- Drive3 Library Loads: 10239 Library Unloads:10234 Drive Loads:1598 MB Written: 10083308 MB Read:5886656 Cleans: 0 -- Drive4 Library Loads: 6877 Library Unloads:6873 Drive Loads:6461 MB Written: 44438571 MB Read:25808491 Cleans: 0 -- Drive5 Library Loads: 6528 Library Unloads:6527 Drive Loads:6705 MB Written: 47722355 MB Read:28614867 Cleans: 0 -- Drive6 Library Loads: 1933 Library Unloads:1932 Drive Loads:2144 MB Written: 11874084 MB Read:6888474 Cleans: 0 -- Drive7 Library Loads: 2051 Library Unloads:2050 Drive Loads:2823 MB Written: 27765506 MB Read:7691746 Cleans: 0 -- Drive8 Library Loads: 2102 Library Unloads:2101 Drive Loads:2265 MB Written: 12299072 MB Read:8002016 Cleans: 0 These tape drives have transferred thousands of terabytes of data (what's that.. exabyte?), apparently without cleanings. What can I check to find out why, other than what I mentioned above (Lib set to AUTO, and drives set in TSM to none). Thanks in advance Todd
Re: BackupSet creativity
I asked my Netware guy about Native File Access, and he said it isn't supported on this Netware version (5.0). Only on 5.1 and up. Thx, though... I hadn't even heard of Native File Access before. It may come in handy at a later date. Bill Boyer [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 09/30/2003 07:58 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: BackupSet creativity How 'bout loading native file access on the Netware volume? I don't remember seeing what version of Netware and if you're at a level that supports this, but then you can access the Netwar volume from the Windows box without a Netware client loaded. Bill Boyer DSS, INc. -Original Message- From: ADSM: Dist Stor Manager [mailto:[EMAIL PROTECTED] Behalf Of Todd Lundstedt Sent: Monday, September 29, 2003 5:34 PM To: [EMAIL PROTECTED] Subject: Re: BackupSet creativity Newp.. This NT server is a locked down NAS server. That was option suggested, investigated, and declined. We will probably just use multiple workstations to do copies. Ted Byrne [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 09/29/2003 03:50 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: BackupSet creativity You're going to have to use that workstation to copy files from point A to point B, after all. Todd, Rather than using a workstation and send the data across the network twice, could you install a Netware client (temporarily) on the NT server, map the drive, and just do a copy from Netware to NT directly? Ted
BackupSet creativity
Although this shouldn't matter, I am running TSM 5.1.7.1 on AIX 4.3.3. We are moving a large number and size of files from a Netware server to an NT server, and I am trying to come up with a way to help the process so we don't have to use a workstation to copy files from a Netware mount to a Windows mount (which really slows things down because the data has to travel from the Netware server to the workstation and then to the NT server all on one network connection). I can't give client access rights to an NT server for Netware backups because of the file/folder rights/properties/attributes issues. So, my bright idea was to generate a backupset of the Netware volume to the TSM server, delete the backup set, define that tape volume to the NT server as a backup set, and restore the data from TSM (my understanding is a backupset is unencrypted backups with no rights/properties information... just files and folders, perhaps readonly, hidden etc attributes included). The generate backupset went well (after I used move nodedata to place all the files on one tape). The delete and define had no noticeable issues. We are trying to do the restore of the backupset to the NT server now, but nothing is moving. The session started only has 4.5K sent and 694bytes rcv'd. It has been sitting like that for 20 minutes. I visually checked the tape drive, and the light is flashing as if the drive is doing something (LTO library). The NT admin verifies that there are no folders/files restored. Was this a pipe dream to begin with?
Re: BackupSet creativity
Oops.. forgot to mention that.. yeah. we did a restore to a specific location, with the default path setting. It does look like it is just reading through the entire tape. This was a backupset of a single filespace, not everything on the server. So I would hope that it is merely file/folders on the tape, and nothing fancy. However (and not to derail my initial plea for help), how can the restore process on my NT server (test box) force a restore TO my Netware server (production box) across the net work? That entire last statement of yours concerns me. I didn't think that was possible, but perhaps I am misinterpreting what you mean there. Ted Byrne [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 09/29/2003 11:52 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: BackupSet creativity Todd, I'm guessing that you will eventually end up with an error message stating that there was no match found, or something to that effect. It may take a long time to get to that point. Backupsets can be very slow to process; you typically wind up reading through the entire tape before the command completes (or errors out). There may also be inherent client compatibility issues that will prevent this from working at all. For one thing, the filespace naming format is completely different between NT and Netware. If you did not specify a destination for the restore, there will be no filespace/drive that matches the source on the server that you are running the restore from. It's always a good idea to specify a destination on a restore like this, anyway - prevents nasty things like overwriting data on the production server across the network, rather than restoring data to the test box that the restore is running on. Ted At 10:23 AM 9/29/2003 -0500, you wrote: Although this shouldn't matter, I am running TSM 5.1.7.1 on AIX 4.3.3. We are moving a large number and size of files from a Netware server to an NT server, and I am trying to come up with a way to help the process so we don't have to use a workstation to copy files from a Netware mount to a Windows mount (which really slows things down because the data has to travel from the Netware server to the workstation and then to the NT server all on one network connection). I can't give client access rights to an NT server for Netware backups because of the file/folder rights/properties/attributes issues. So, my bright idea was to generate a backupset of the Netware volume to the TSM server, delete the backup set, define that tape volume to the NT server as a backup set, and restore the data from TSM (my understanding is a backupset is unencrypted backups with no rights/properties information... just files and folders, perhaps readonly, hidden etc attributes included). The generate backupset went well (after I used move nodedata to place all the files on one tape). The delete and define had no noticeable issues. We are trying to do the restore of the backupset to the NT server now, but nothing is moving. The session started only has 4.5K sent and 694bytes rcv'd. It has been sitting like that for 20 minutes. I visually checked the tape drive, and the light is flashing as if the drive is doing something (LTO library). The NT admin verifies that there are no folders/files restored. Was this a pipe dream to begin with?
Re: BackupSet creativity
Newp.. This NT server is a locked down NAS server. That was option suggested, investigated, and declined. We will probably just use multiple workstations to do copies. Ted Byrne [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 09/29/2003 03:50 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: BackupSet creativity You're going to have to use that workstation to copy files from point A to point B, after all. Todd, Rather than using a workstation and send the data across the network twice, could you install a Netware client (temporarily) on the NT server, map the drive, and just do a copy from Netware to NT directly? Ted
Re: SQL server backups
Yes, but you could stripe the full backup across multiple tape drives (if you have more than one tape drive). This spreads the database backup across multiple tapes, and makes reclamation of those files less problematic. Note: if you stripe a backup, you *must* send the backup to a collocate=filespace storage pool. Otherwise, reclamation could put more than one stripe for a particular backup on a tape, making restoration impossible. Additionally, TDP will allow you to backup differential backups of the database, and even transactions logs, directly to TSM storage. This would cut down backup time (but increase restore time). Hope that helps. Todd ps.. we are backing up an 800+ GB SQL database using stripe=3 to LTO1 drives via SAN in 3-4 hours. pps.. if you implement a FULL with DIFF backup from TDP, make sure your database is not getting dumped to disk at all, or it invalidates the FULL on tape, but TSM is unaware of it. Thomas Denier [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 09/23/2003 12:39 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:SQL server backups My company has a Windows server supporting a very large SQL Server database. Before the last hardware upgrade, the database was backed up with the SQL Server Connect Agent. Since the upgrade the database has been dumped to a flat file which is then backed up by the backup/archive client. Either way the backup arrives at the TSM server as a single file containing about 20 gigabytes of data. A single file this size causes a number of problems. The recovery log gets pinned for hours. Tape reclamation tends to perform poorly. The system administrator is now considering installing TDP for SQL Server. Would this software still send the backup as a single huge file?
Re: SQL server backups
One option would be to redirect those backups to storage pools that don't reclaim. Simply let the retention expire the data, and empty the volumes. It is what I had to do to prevent reclaiming three 280+GB striped files. Thomas Denier [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 09/23/2003 01:11 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: SQL server backups My company has a Windows server supporting a very large SQL Server database. Before the last hardware upgrade, the database was backed up with the SQL Server Connect Agent. Since the upgrade the database has been dumped to a flat file which is then backed up by the backup/archive client. Either way the backup arrives at the TSM server as a single file containing about 20 gigabytes of data. A single file this size causes a number of problems. The recovery log gets pinned for hours. Tape reclamation tends to perform poorly. The system administrator is now considering installing TDP for SQL Server. Would this software still send the backup as a single huge file? Without product, level, and tape drive details, it's hard to give a lot of advice; but it sounds like you have significant tape drive technology issues slowing everything down. Large files are usually the optimal type of objects for backup systems, allowing streaming, minimal tracking updates, etc. I would not approach TDP until you address underlying issues. I'd begin by getting benchmark numbers on your tape drive and other subsystem throughput rates to isolate the bottleneck and address that. I didn't give details about the tape configuration because it is clearly not the problem. We used to be able to back up the database to a disk storage pool initially (the database is slowly but steadily growing) and had the same kind of log pinning issues. The worst form of the tape reclamation pathology occurs with offsite tapes. We tend to get one of the huge files starting at or near the beginning of a 3590 J volume, filling the rest of that volume, and spilling on to the first few percent of another volume. The next offsite reclamation process regards the second volume as a prime candidate for reclamation and recopies the entire file to two new volumes. In some cases we have seen this phenomenon repeatedly for several days running. Onsite tape reclamation handles big spanned files more gracefully, but even there such files result in a poor trade-off between the amount of data movement and the amount of free space generated. The pathalogical behavior results from the relationship between file size and volume capacity, and has absolutely no relation to streaming, tracking updates, or whatever.
Re: Deleting a copypool
Return the tapes to your library (if you have one) and update each volume to readonly or readwrite. They will be deleted from the storage pool, and then you can delete the storage pool. [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 09/23/2003 02:21 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Deleting a copypool Hi, I created a copypool a month ago so I could copy a set of tapes to be stored offsite. Now, I no longer need those tapes. Reading the help manual, I first deleted each tape with 'DISCARDDATA=YES' option. Every tape was successfully deleted. I then proceeded to delete the copy pool. However, TSM refused to remove the pool, saying it still containing tapes. Can someone help me to get those tapes out of this copypool? The access of those tapes are all marked 'OFFSITE' in within this pool right now. My TSM server version is 5.1.6.2 running on AIX 5.1 TIA
Re: Very long Netware restore
MTU=Maximum Transmission Unit and is a parameter of the network card/driver. Kamp, Bruce [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 09/23/2003 04:43 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: Very long Netware restore I don't the exactly what MTU means Basicly it is the packet size. We found this out after my network guys ran Sniffer on our network. The difference in restore tmes was 2-3 days to about 10-12 hours! Your network people should be able to give a better definition! I colocated my tapepool after my first big reatore. --- Bruce Kamp Midrange Systems Analyst II Memorial Healthcare System E-mail: [EMAIL PROTECTED] Phone: (954) 987-2020 x4597 Fax: (954) 985-1404 --- -Original Message- From: William Rosette [mailto:[EMAIL PROTECTED] Sent: Tuesday, September 23, 2003 3:35 PM To: [EMAIL PROTECTED] Hey Bruce, I had a similar problem with 500 GB of info on 2 clients of Novell/Netware, 1 200 GB and the other 300 GB. It also took days. I was told to collocate and restore the dirs only first. What does MTU stand for? Is it a TSM client setting?, server setting? or Novell/Netware setting? Thank You, Bill Rosette Data Center/IS/Papa Johns International WWJD Kamp, Bruce [EMAIL PROTECTED] To: [EMAIL PROTECTED] Sent by: ADSM: cc: Dist StorSubject: Re: Very long Netware restore Manager [EMAIL PROTECTED] .EDU 09/23/2003 03:19 PM Please respond to ADSM: Dist Stor Manager Had this problem before! It was an MTU size problem. Once that was fixed the restore flew! -- Bruce Kamp Midrange Systems Analyst II Memorial Healthcare System E: [EMAIL PROTECTED] P: (954) 987-2020 x4597 F: (954) 985-1404 --- -Original Message- From: Richard Rhodes [mailto:[EMAIL PROTECTED] Sent: Tuesday, September 23, 2003 2:07 PM To: [EMAIL PROTECTED] We recently did a full restore of a Netware server. Basically, we're interested if the restore time sounds reasonable - it doesn't to us. The server that was restored is actually at a remote site from the TSM server location, but the restore was done at the TSM site, then, driven to the remote site for installation. The restore specs read like this: Compression: Files compressed using TSM client compression Size: 240gb Files: 938,617 Restore Time: 50 hours Throughput: around 1.3mb/s average speed (240gb/50hr) Network: 100MB/fdx, local to TSM server (admins checked for duplex mismatch) Tape Drives: IBM 3590 Netware Server Version: NW5.1 SP3 Dell 2650 1 - 2ghz xeon processor 2gb memory 513mb cache on processor This just doesn't sound right. After this was all done, we created a backupset to see just how fast TSM could access the servers files. Backupset Creation Time: 11 hours From Tape Drive: IBM 3590 To Tape Drive: IBM 3590 So . . . . the bottleneck doesn't appear to be the TSM server. Any thoughts as to why our restore took 50 hours??? The obvious answer is that the files had to be uncompressed on the client, but I would have thought a 2ghz processor would be able to uncompress much more than a 1.3mb/s data stream. Thanks Rick - The information contained in this message is intended only for the personal and confidential use of the recipient(s) named above. If the reader of this message is not the intended recipient or an agent responsible for delivering it to the intended recipient, you are hereby notified that you have received this document in error and that any review, dissemination, distribution, or copying of this message is strictly prohibited. If you have received this communication in error, please notify us immediately, and delete the original message.
Move NodeData, and copy storage pools
TSM 5.1.7.1 on AIX 4.3.3 I have moved data for several nodes from one primary storage pool to another primary storage pool. Each of these primaries has a different copy storage pool defined. OrigPrimaryStg -- OrigCopyStg NewPrimaryStg -- NewCopyStg move nodedata node1,node2 from=OrigPrimaryStg to=NewPrimaryStg After the move, the backup of the NewPrimaryStg copied the data to NewCopyStg, but how do I remove the data from OrigCopyStg? Per the documentation, one cannot move data from one copy storage pool to another. Am I stuck with doing a move nodedata within the OrigCopyStg, then deleting the volumes? I am assuming after the move nodedata, I could delete vol xx discarddata=yes for the OrigCopyStg volumes that contain ONLY data from node1 and node2, and it will NOT delete the data in the primary storage pool (NewPrimaryStg) and it's copy storage pool (NewCopyStg). Is that a correct assumption, too? Help on Delete Volume indicates copies will be deleted when primaries volumes are deleted, but it doesn't specify if primaries are deleted when copy volumes are deleted. Thanks in advance Todd
Re: Open File Support v5.2.0.1
I had my first experience with 5.2 client code a few weeks ago. We had initially manually selected to install the open file module (among others) when we installed the client on a WinXP machine. I received the same errors you listed. After much trial and error, we had ran the setup again, chose to modify the install, and removed the open file module. After that, backups went normally. Todd Joshua S. Bassi [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 08/05/2003 01:54 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: Open File Support v5.2.0.1 Did you install to a machine with an existing version of TSM? If so, was the TSM LVSA module installed (i.e. were you enabled for online image backup)? This was a brand new machine with no TSM code on it. How *exactly* did you perform the install of the client? If you chose to install all components, open file support is included. The standard installation method was used, just default modules were loaded. I don't follow the part that says ... it tries to do an image backpu of the filesystem Are you saying that when you attempt to run dsmc incremental it instead runs dsmc backup image? The error message we get is: ANS1228E Sending of object \\hostname\c$ failed ANS1381E The snapshot cache location cannot be located on the same volume that is being backed up. ANS1228E Sending of object \\hostname\c$ failed ANS1377E Unable to perform operation using a point-in-time copy of the filesystem on '\\hostname\c$. The backup/archive operation will continue without snapshot support. I am not 100% sure what it is doing right now and am hoping that you can shed some light on the situation. --- Joshua S. Bassi IBM Certified - AIX/HACMP, SAN, Shark Tivoli Certified Consultant- ADSM/TSM [EMAIL PROTECTED] Cell (831) 595-3962
Internal Server Error Detected
We just upgrade our server version from 4.2.1.7 to 5.1.7.1 on AIX 4.3.3. We had to upgrade to stay in support. The client having problems is 3.1.0.8 running on Netware 4.10 (I understand Netware 4.10 cannot use TSM B/A Clients higher than 3.1.0.8). I understand B/A 3.x is way out of support. This has been happening since the upgrade of the server code. I have a DIRMC set in a client option set assigned to this node, which points to sequential file based volumes in that management class. The rest of the files backup to our default management class, which points to random disk volumes. When we attempt to backup files and directories, we get the following in the activity logs... 08/05/03 12:20:59 ANR0406I Session 27441 started for node SFRMC_SURGERY (NetWare) (Tcp/Ip x.x.x.x(3736)). 08/05/03 12:21:08 ANR8340I FILE volume /tsmdirdisk/1EEB.BFS mounted. 08/05/03 12:21:08 ANRD bfcreate.c(1918): ThreadId64 Destination switched from L1_DIR_FILEPOOL to L1_PRI_DSK in the middle of a transaction. 08/05/03 12:21:08 ANR0530W Transaction failed for session 27441 for node SFRMC_SURGERY (NetWare) - internal server error detected. I did some searching on the net, using the text of the ANRD message, and found one post where removing the DIRMC setting helped. That didn't help for us. (Side note: First, our DIRMC setting was set via a client option set, and I couldn't quickly see how to remove a client option set value from a specific node, so I copied the existing client option set to a new name, and deleted the option for DIRMC in the new set, and assigned the node to that new client option set. That didn't help, so I renamed the node in TSM, created a new node with no client option set value, and tried the backup again. Same error. Of course, we restarted the NLMs after every change to be certain.) Has anyone found a way to get a Netware 4.10 server using B/A client 3.1.0.8 to work with TSM Server 5.1+?
Re: Internal Server Error Detected
Well, this is puzzling indeed. It works for you for a WinNT B/A client at 3.1.0.8 with the DIRMC option removed, and it worked for the Netware user who's post I found (also removing the DIRMC option), yet it doesn't work for me with or without the DIRMC option. We are already looking at other options, but it would be great if we didn't have to purchase another backup system for just these two nodes. I am not giving up on this. If anyone else has ideas, please let me know. Thanks. Todd Tab Trepagnier [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 08/05/2003 05:07 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: Internal Server Error Detected Todd, I saw the same thing with client V 3.1.0.8 on Windows NT sending data to TSM server 5.1. I, too, use the DIRMC option on my production TSM system. On my test system, the V 3.1.0.8 client worked OK with V 5.1 server without the DIRMC option implemented. There is no workaround that I know of other than a client upgrade or removing the DIRMC option. Our V 4.1 clients didn't exhibit the problem. We used the V 5.1 upgrade to get the clients, too, so now almost all of our clients are running 5.1. We had been running a mix of 3.1, 4.1, and 4.2 clients. Tab Trepagnier TSM Administrator Laitram LLC Todd Lundstedt [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 08/05/2003 12:43 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Subject:Internal Server Error Detected We just upgrade our server version from 4.2.1.7 to 5.1.7.1 on AIX 4.3.3. We had to upgrade to stay in support. The client having problems is 3.1.0.8 running on Netware 4.10 (I understand Netware 4.10 cannot use TSM B/A Clients higher than 3.1.0.8). I understand B/A 3.x is way out of support. This has been happening since the upgrade of the server code. I have a DIRMC set in a client option set assigned to this node, which points to sequential file based volumes in that management class. The rest of the files backup to our default management class, which points to random disk volumes. When we attempt to backup files and directories, we get the following in the activity logs... 08/05/03 12:20:59 ANR0406I Session 27441 started for node SFRMC_SURGERY (NetWare) (Tcp/Ip x.x.x.x(3736)). 08/05/03 12:21:08 ANR8340I FILE volume /tsmdirdisk/1EEB.BFS mounted. 08/05/03 12:21:08 ANRD bfcreate.c(1918): ThreadId64 Destination switched from L1_DIR_FILEPOOL to L1_PRI_DSK in the middle of a transaction. 08/05/03 12:21:08 ANR0530W Transaction failed for session 27441 for node SFRMC_SURGERY (NetWare) - internal server error detected. I did some searching on the net, using the text of the ANRD message, and found one post where removing the DIRMC setting helped. That didn't help for us. (Side note: First, our DIRMC setting was set via a client option set, and I couldn't quickly see how to remove a client option set value from a specific node, so I copied the existing client option set to a new name, and deleted the option for DIRMC in the new set, and assigned the node to that new client option set. That didn't help, so I renamed the node in TSM, created a new node with no client option set value, and tried the backup again. Same error. Of course, we restarted the NLMs after every change to be certain.) Has anyone found a way to get a Netware 4.10 server using B/A client 3.1.0.8 to work with TSM Server 5.1+?
Re: Internal Server Error Detected
I can't speak to how different NW 4.10 and 4.11 are. Our Netware guy did some research, and found that version TSM BA 3.7.x.x (can't remember the x's) supported NW 3.1, 3.2, 4.11, and some other higher versions, but it skipped 4.10. TSM BA 3.1.0.8 does specify NW 4.10. When we installed TSM, starting on 4.2.1.7, we couldn't get any of the 4.2 clients to work on the NW 4.10 servers. The consultant found the 3.1.0.8 version, and it worked (unsupported, but it worked). One of the NW servers is running an application we expect to have replaced by Oct 1, so we should be able to decommission that NW410 server by Jan1 (must run parallel). The other server we are discussing upgrading/migrating it with the team responsible for it. Until then, we are using some other workaround we have devised to get the backups we need. Shannon Bach [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 08/07/2003 10:27 AM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: Internal Server Error Detected All our Netware servers are now at V6 but it took at least a year to get them all upgraded. In the meantime all our Netware V4.11 boxes were using the TSM Client version 4.1.3.0 and had no problems with backups. Is NW V4.10 that much different? And if so, wouldn't it be easier to upgrade your NW's to V4.11? I know next to nothing about NetWare except as a big chunk of my TSM Backup Clients so I may be way off base here, but it seems to me that it is much easier to upgrade the NW boxes one level than to consider bringing in another backup product that will not be half as versatile as TSM. Shannon Bach Madison Gas Electric Co. Operations Analyst - Data Center Services Office 608-252-7260 Fax 608-252-7098 e-mail [EMAIL PROTECTED]
Re: Reclamation of Copy Dirs
I ran into this exact problem when I had one 1GB random access DISK based primary storage pool (about 27% used). I didn't backup it up to it's own copy storage pool, but I did back it up to the same copy storage pool where I copied the other management classes. When reclamation occurred, I would often have one or two offsite tapes that were 0% full, 99% reclaimable, and the reclaim job ran all day with one output, and no input volumes mounted. To fix the immediate needs, I would find a nice non-stormy day and update reclamation=100, cancel existing reclamation, go offsite, retrieve the offending tapes, insert them to the library, update them to readonly, and move the data manually. Those moves went fairly quickly... lightning compared to the reclamation. To fix the long term needs, when I had time to do it, I created a new device class type of DIR_FILE with an estimated capacity of 10MB, then created a new sequential access storage pool using that DIR_FILE device class that allowed up to 100 volumes (approx 1GB total), then pointed the DIRMC to use that new storage pool. Then I moved the data from the random disk volumes to the new storage pool (so it filled up just over 27 volumes right away). After all that, the reclamation is going at a better rate. I think it is because the one 1GB disk volume containing all of the directory objects was just too resource intensive to scan. The 10MB volumes are much faster to scan through. Personally, it doesn't make sense that a random access volume would have that much trouble finding the file, but I had read other posts with problems similar to mine, so I made those changes. Hope this helps Todd Debi Randolph [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 08/05/2003 12:44 PM Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Fax to: Subject:Re: Reclamation of Copy Dirs Yes, it's forever and 2 days with no tape drive contention. We're using 3590E1A's, but moving to LTO in the next 30 to 60 days. I only have 12 volumes utilized in the Copy Dirs pool. So we've turned over the specific volumes a number of times. So I don't think it's any one volume that could be causing the problem, as far as read/write errors go. How does reclamation on this pool work? When you reclaim Copy Dirs, it only uses one tape drive, just an output volume, no input volume. All the other storage pools use two drives, an input volume and an output volume. So is it going back to the client to re-write the data to a new tape? Thanks, Deb Randolph Richard Sims [EMAIL PROTECTED] To: [EMAIL PROTECTED] Sent by: ADSM: cc: Dist StorSubject: Re: Reclaimation of Copy Dirs Manager [EMAIL PROTECTED] .EDU 08/05/2003 11:17 AM Please respond to ADSM: Dist Stor Manager Is it supposed to take forever and 2 days to reclaim my Copy Dirs pool? Not for the price of the product... You haven't told us if you are checking for tape drive contention, as in drives busy with higher priority tasks. If no contention, you may have some difficult tapes, consuming costly retries in reading or writing. To try to isolate the problem, you can do Move Data on indicative tapes and follow the progress, watching for particular points where things crawl, and go from there, as in examining I/O load or paging on your server. We also don't know your tape technology type: as you may have seen from historic posting, start-stop on some drives is painful. Richard Sims, BU ***N O T I C E*** The information contained in this e-mail, and in any accompanying documents, may constitute confidential and/or legally privileged information. The information is intended only for use by the designated recipient. If you are not the intended recipient (or responsible for the delivery of the message to the intended recipient), you are hereby notified that any dissemination, distribution, copying, or other use of, or taking of any action in reliance on this e-mail is strictly prohibited. If you have received this e-mail communication in error, please notify the sender immediately and delete the message from your system. *** ***N O T I C E*** The information contained in this e-mail, and in any accompanying documents, may constitute confidential and/or legally privileged information. The information is intended only for use by the designated recipient. If you are not the intended recipient (or responsible for the delivery of the message to the intended recipient), you are hereby notified that any
Version 5 features
I could swear I read somewhere that TSM v5 added a feature to client schedules that allowed for Pre- and Post-schedule commands. I can't remember where I read that, nor can I find the information about it in the version we have installed now (just upgraded from 4.2.1.7 to 5.1.7.1). Does v5 allow this? Where do I read about it? Thanks in advance. Todd
Netware backup oddities
Server is AIX 4.3.3 running TSM 4.2.1.7 Prior to this weekend, everything seemed to be fine on this Netware 4.11 server running B/A Client 4.1.3.0. Compression is set to ON at the Node Definition level. It was inspecting its 111K+ files and backing up between 200-1000 files a day (about 60-200MB) and completing that entire processes within 20 or 30 minutes, according to log information. Friday night, the backup didn't complete until Sunday morning, running 31+ hours, backing up 18K+ files (30GB+). Sunday night's backup was still running Monday morning when we came in. We cycled that Netware server (which killed the backup in progress). We even cycled the TSM server process and OS (as a request from TSM support regarding another issue). I took the opportunity to increase the log retention days, and turn on verbose logging. (On the sessions where the backup amounts increased, the dsmerror.log indicated several instances of session lost due to inactivity, but that is quite possible as some of the files on this server are quite large, and compression could have caused a session to time out.) Last night's backup was still running when we came in this morning. With the verbose logging, I could see what files were being backed up. I started a GUI restore session with view active/inactive set on, and navigated to some of the directories that I noticed in the logs. In all cases that I checked... 1. The entire directory and files within had been in existence since before TSM was installed on this node (Nov, 01). 2. Those files in that directory had backed up originally on day one of TSM on this node 3. Some of the files had recently been backed up again early in Feb 03, even though there was no evidence of them changing. 4. Those same files started backing up again this weekend, again, even though there was no evidence of the changing. They appear to be backing up every session, even though the session on Friday night - Sunday morning successfully completed, and they backed up then. One more thing, we do have the NWIGNORECOMPRESSBIT flag set to YES, but that really shouldn't have mattered as many of the files that I noticed should have been Netware compressed for a very very long time, and would not have been accessed by a user. Any clues? Thanks, Todd
Re: Cancel All Processes
I have a script that cancels any migration, reclamation, move, and backup stgpool processes. I modified it slightly here, but it should work to cancel any process running in TSM. However, remember that the process will not necessarily cancel immediately. It has to finish what it is doing at that time before canceling, for instance a 150GB file that just started reclaiming will take approximately 3 hours to reclaim (on my system). A cancel process will not actually cancel that process until that file is finished moving. Anyway.. here is the slightly modified script. PROCESSES=somepath/procs dsmadmc -id=someTSMid -pass=TSMidpasswd q pro | awk '{ print $1 }' $PROCESSES for ProcNo in `cat $PROCESSES` do dsmadmc -id=someTSMid -pass=TSMidpasswd cancel pro $ProcNo done I haven't tested this, but it should work, maybe with some slight modification to the 'awk' parameters. |+ || Dearman, | || Richard | || [EMAIL PROTECTED]| || DU | || Sent by: | || ADSM: Dist | || Stor Manager | || [EMAIL PROTECTED]| || IST.EDU | ||| ||| || 03/25/2003| || 03:57 PM | || Please respond| || to ADSM: Dist| || Stor Manager | ||| |+ | | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Cancel All Processes | | Does anyone have a script to cancel all process? For example you can cancel sess all to cancel all sessions but there is no cancel proc all command. So there must be a scripting way of doing this. THanks ***EMAIL DISCLAIMER*** This email and any files transmitted with it may be confidential and are intended solely for the use of th individual or entity to whom they are addressed. If you are not the intended recipient or the individual responsible for delivering the e-mail to the intended recipient, any disclosure, copying, distribution or any action taken or omitted to be taken in reliance on it, is strictly prohibited. If you have received this e-mail in error, please delete it and notify the sender or contact Health Information Management 312.996.3941.
Re: Selective backups
Thanks Scott. We are setup using POLLING, and our Randomization is set to 10% with a Client Action Duration of 5 days. Perhaps this is why it ran so much later. The startup window is 720 minutes (12 hours). I can change my Client Action Duration to 1 day, that will get me a start window of 144 minutes, run the schedule, then change it back. Are there any other gotchas that might hit me there? I placed commas without spaces in between the objects because that's what the Help indicated to do. I am guessing I would need the double quotes around the list of drives if I am using spaces. Correct? As far as why I need to do this... I thought I needed to do a selective backup of all directories on all servers so that the ones that are stored on tape could be backed up again, and then stored on disk only, eventually allowing the ones on tape to expire off. We do have a few Netware servers, and a large number of NT/W2K servers. So, in order to make sure we have everything moved to a the new management class, I thought the selective backup would be warranted. |+ || Scott McCambly| || [EMAIL PROTECTED]| || ANADA.CA | || Sent by: | || ADSM: Dist | || Stor Manager | || [EMAIL PROTECTED]| || IST.EDU | ||| ||| || 03/18/2003| || 10:00 AM | || Please respond| || to ADSM: Dist| || Stor Manager | ||| |+ -| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: Selective backups | -| I suspect your immediate action ran 7 hours later because your node is using POLLING schedmode and you have randomization set to 50%. Immediate action schedules are defined with a huge starting window, so although I don't typically use POLLING, I've never understood how it would ever work for those nodes (should be called eventual action schedule :-) ). The job failed because of the commas in your objects list. Just type it exactly as you would on the command line. And all you need is C:\ D:\ G:\ - the asterix is not necessary, but shouldn't hurt. Most importantly however is why you think you need to do this in the first place. Unless you have excessively large ACL's on your directories, they are stored entirely in the TSM database. They didn't used to many versions ago, so what you are setting up used to be valid. If this is your situation, then great! I'm assuming that most people today only use DIRMC to avoid binding to a management class that has NOLIMIT for object retonly. Does anyone else out there use DIRMC? Todd Lundstedt wrote: I have recently setup a Directory Management Class, some server-side Client Option Sets pointing DIRMC to that management class, and assigned those option sets to nodes. So now, all new/changed directories should be being managed by that management class (disk based storage, no migration, etc.). I can see that going to each node and issuing a dsmc selective -dirsonly -subdir=yes c:\* g:\* and any other drives that exist will re-backup versions of all directories on the listed disk drives. But, that's a lot of work. So I tried an immediate action schedule on one of the servers. Action = Selective Options = -dirsonly -subdir=yes Objects = c:\*,d:\*,g:\* and that immediate action schedule finally tried to run 7 hours later, and didn't work. Is there anything I can do from the server to kick off a -dirsonly -subdir=yes backup of ALL existing drives in a list of nodes? From what I am gathering, the selective backup requires a filespace and I have not been able to get *:\* to work yet. Or, do I need to even worry about this? My intention was to (eventually) get all backup versions of directory objects off of tape and stored only on disk. Thanks in advance.
Re: Selective backups
Thanks Andy. Here is what the Help web page shows me when I click on the Help button in the Admin GUI for Immediate Action Schedules... ++ Objects Enter the objects for which the specified action is performed. * You can specify more than one object by separating the objects with commas, with no intervening spaces. * etc. ++ Those two quotes seem to contradict one another. But I will try following your quote (Scott's suggestion to give credit where credit is due), since following the instructions I read didn't work. Also, Andy, do you think there is any gotchas I need to be concerned with if I set my Client Action Duration via the Immediate Action Schedule GUI from the current five (days) to one? I would set it back again after all the clients in the list had a chance to run (which should be withing 2.5 hours if I am doing my math correctly (1440 minutes (one day) * 10% randomization = 144 minutes). Thanks in advance. Todd |+ || Andrew Raibeck| || [EMAIL PROTECTED]| || M.COM| || Sent by: | || ADSM: Dist | || Stor Manager | || [EMAIL PROTECTED]| || IST.EDU | ||| ||| || 03/18/2003| || 11:22 AM | || Please respond| || to ADSM: Dist| || Stor Manager | ||| |+ -| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: Selective backups | -| Here is what HELP UPDATE SCHEDULE or HELP DEFINE SCHEDULE shows me for the OBJECTS parameter: OBJects Specifies the objects for which the specified action is performed. Use a single space between each object. etc. And yes, you'll need to put quotes around the list of objects (this is a general rule for any parameter that contains blank spaces, as described in the Command Entry Rules in the server admin reference). The best way to think of this is to enter the OBJECTS and OPTIONS strings the same way you would enter them from the client command line. For example, a client command like this: dsmc archive c:\dir1\ d:\dir2\ f:\dir3\ -subdir=yes -archmc=mymc would be scheduled like this: def sch standard mysched action=archive objects=c:\dir1\ d:\dir2\ f:\dir3\ options=-subdir=yes -archmc=mymc ... Regards, Andy Andy Raibeck IBM Software Group Tivoli Storage Manager Client Development Internal Notes e-mail: Andrew Raibeck/Tucson/[EMAIL PROTECTED] Internet e-mail: [EMAIL PROTECTED] (change eye to i to reply) The only dumb question is the one that goes unasked. The command line is your friend. Good enough is the enemy of excellence. Todd Lundstedt [EMAIL PROTECTED] Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] 03/18/2003 09:53 Please respond to ADSM: Dist Stor Manager To: [EMAIL PROTECTED] cc: Subject:Re: Selective backups Thanks Scott. We are setup using POLLING, and our Randomization is set to 10% with a Client Action Duration of 5 days. Perhaps this is why it ran so much later. The startup window is 720 minutes (12 hours). I can change my Client Action Duration to 1 day, that will get me a start window of 144 minutes, run the schedule, then change it back. Are there any other gotchas that might hit me there? I placed commas without spaces in between the objects because that's what the Help indicated to do. I am guessing I would need the double quotes around the list of drives if I am using spaces. Correct? As far as why I need to do this... I thought I needed to do a selective backup of all directories on all servers so that the ones that are stored on tape could be backed up again, and then stored on disk
Selective backups
I have recently setup a Directory Management Class, some server-side Client Option Sets pointing DIRMC to that management class, and assigned those option sets to nodes. So now, all new/changed directories should be being managed by that management class (disk based storage, no migration, etc.). I can see that going to each node and issuing a dsmc selective -dirsonly -subdir=yes c:\* g:\* and any other drives that exist will re-backup versions of all directories on the listed disk drives. But, that's a lot of work. So I tried an immediate action schedule on one of the servers. Action = Selective Options = -dirsonly -subdir=yes Objects = c:\*,d:\*,g:\* and that immediate action schedule finally tried to run 7 hours later, and didn't work. Is there anything I can do from the server to kick off a -dirsonly -subdir=yes backup of ALL existing drives in a list of nodes? From what I am gathering, the selective backup requires a filespace and I have not been able to get *:\* to work yet. Or, do I need to even worry about this? My intention was to (eventually) get all backup versions of directory objects off of tape and stored only on disk. Thanks in advance.
Re: Renaming a W2K node
Thanks for everyone's responses on this. With the suggestions given here, I was able to write up a procedure to assist our NT/W2K admins with this function. It involves using the DSMCUTIL command instead of the wizard, and a little thought before doing it. In these special cases, I wanted the TSM node name to be something other than the NETBIOS name, and the scripts we have setup for the NT admins to run includes registering the node, and installing the services with %computername%. Which means, the filespace names will default to \\%computername%\driveletter$, for a totally different named node. However, if the NETBIOS name is changing along with the nodename, the filespace names will need to be changed as well (or else they will just backup full again). Anyway.. Thanks again everyone. Todd |+-- || Prather, Wanda| || [EMAIL PROTECTED]| || HUAPL.EDU | || Sent by: ADSM: | || Dist Stor | || Manager| || [EMAIL PROTECTED]| || T.EDU | || | || | || 03/13/2003 01:13| || PM | || Please respond | || to ADSM: Dist | || Stor Manager | || | |+-- -| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: Renaming a W2K node | -| Everything you did is OK. And with old versions of the client (V3.1, for instance) that is all you had to do. But on W2K, you need more steps for the service. 1) Repeat all the stuff you did, that makes the GUI OK. 2) Start the GUI, pull down Utilities, Setup Wizard 3) Click (only) the check box for configure scheduler. - NEXT 4) Click modify existing scheduler - NEXT 5) On the next page, click to highlight the service name - NEXT 6) Keep walking through and take all the defaults except put in the the new nodename and password. 7) Click FINISH. 8) restart the scheduler service Bottom line is that the service installation saves stuff in the registry and you have to fix it this way. If this doesn't fix it, I would try using the setup wizard to uninstall the old service, then reinstall it with the new nodename. -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED] Sent: Wednesday, March 05, 2003 7:00 PM To: [EMAIL PROTECTED] Subject: Renaming a W2K node AIX 4.3.3 running TSM Server 4.2.1.7. I have a few client nodes that are W2K servers. Oddly, they are named after our NT servers (NT-ServerA, NT-ServerB, etc..). The dsm.opt file contains the NT machine name as the NodeName. I would like to change the NodeName to W2K-ServerA, but not rename the machine name. ... I stopped the scheduler service on the client node. ... I modified the name in dsm.opt, and saved the file ... I renamed the node on the TSM Server (rename node nt-servera w2k-servera, or whatever the command is... I actually used the web admin GUI to do it). ... I entered dsmc query session at the command line of the client node. ... ... it indicated it was node W2K-ServerA, requested the user ID, and password.. I entered (defaulted) the ID, and keyed in the password for that node. The information was returned as expected. ... I entered dsmc query session at the command line of the client node, again. ... ... The information was returned as expected, without ID and password being required (by the way, the opt file does have passwordaccess generate), still indicating it was the new nodename, W2K-ServerA. ... I started the service up... it started and quietly failed (no GUI messages on the screen). I checked the dsmsched.log file, and it appeared that it attempted to connect using NT-ServerA instead of W2K-ServerA. Baffled.. so ... I checked to properties on the service.. I didn't see anything that indicated NT-ServerA. ... I started the Backup/Archive GUI (vers 5.1.x) and checked the preferences there. Everything was as I would have expected it to be based
Client Connection Issues
TSM server 4.2.1.7 running on AIX 4.3.3 TSM Client 5.1.5.0 running on Netware 5 Start a Backup Archive GUI session from the Client Node list (click on the Link to Node icon). The Java GUI loads (if this is the first time accessing a 5.x client from that workstation, the Java install runs without issue). Click on the Backup or Restore or Archive or Retrieve button, and nothing happens. You can see the GUI button get pressed, but it never un-presses, and the appropriate utility never starts. The main Java GUI goes in to a state where the graphical portions won't refresh if another application is placed on top of it, or if it is minimized then restored. This worked fine last week, so I know everything was working alright. Other clients' v5 GUIs work as expected. Any clues where to look? Thanks in advance
Renaming a W2K node
AIX 4.3.3 running TSM Server 4.2.1.7. I have a few client nodes that are W2K servers. Oddly, they are named after our NT servers (NT-ServerA, NT-ServerB, etc..). The dsm.opt file contains the NT machine name as the NodeName. I would like to change the NodeName to W2K-ServerA, but not rename the machine name. ... I stopped the scheduler service on the client node. ... I modified the name in dsm.opt, and saved the file ... I renamed the node on the TSM Server (rename node nt-servera w2k-servera, or whatever the command is... I actually used the web admin GUI to do it). ... I entered dsmc query session at the command line of the client node. ... ... it indicated it was node W2K-ServerA, requested the user ID, and password.. I entered (defaulted) the ID, and keyed in the password for that node. The information was returned as expected. ... I entered dsmc query session at the command line of the client node, again. ... ... The information was returned as expected, without ID and password being required (by the way, the opt file does have passwordaccess generate), still indicating it was the new nodename, W2K-ServerA. ... I started the service up... it started and quietly failed (no GUI messages on the screen). I checked the dsmsched.log file, and it appeared that it attempted to connect using NT-ServerA instead of W2K-ServerA. Baffled.. so ... I checked to properties on the service.. I didn't see anything that indicated NT-ServerA. ... I started the Backup/Archive GUI (vers 5.1.x) and checked the preferences there. Everything was as I would have expected it to be based on the values in the dsm.opt file, including the node name being W2K-ServerA After much searching of documentation and scratching of head, I punted and changed everything (dsm.opt, nodename on TSM server, etc.) back to NT-ServerA and the service started up and stayed up. What am I missing to change the node name? Something is still using the old node name when it attempts to contact the TSM server, and of course, that nodename no longer exists on the TSM server. I did do some looking around in that quickFaQ that Richard links to all the time... didn't find anything in there that jumped out at me. Same with the last 3 months of ADSM.org archives, and IBM's horridly slow and completely non-user-friendly support pages. I am lost... Thanks in advance Todd
Backup Strategy and Performance using a PACS Digital Imaging System
Heya *SMers, This is a long one, grab a beverage. I am running TSM 4.2.1.7 on AIX 4.3.3 on a dual processor B80 backing up around 112 nodes with 5.2TB capacity 2.3TB used, storing 18.6 million objects (3.2TB) in the primary and copy storage pools. The TSM database is 8.7GB running around 80% utilization. All but a few of the nodes are server class machines. We backup about 250GB a night, and on weekends we do full TDP SQL database backups to the tune of an additional 450 GB (and growing). Expiration processing occurs daily, and completes in approximately 70 minutes. The four Fuji PACS servers we have are included in the above numbers, but only the application and OS, not the images and clinical notes (less than 1k text files). FYI.. where TSM and disk management are concerned, Fuji is the DEVIL!. Each image, and each 1k note file with text referencing an image are stored in their own directory.. image_directory/imagefile and text_directory/textfile.. a one to one relationship. To backup the directories/textfiles now takes the backup process over 12 hours to complete, incrementally backing up very little. The backup has to scan the files to see what needs to be backed up (this is not TSM yet, but some other backup software). The powers that be are asking what it would take to move all of the data stored on DVDs in the DVD jukebox (images) to magnetic media disk based storage. Then, start backing all of that up to TSM. I have some numbers from the PACS administrator. On the four PACS servers, the additional data they would like TSM to backup tallies up to... 1.5+ million folders 1.0+ million files (yes... more folders than files...) 2.2+ TB storage (images and text files) All of this data will not change. Once it backs up, it will very likely never need to be backed up again. Because of that, I am recommending three tape storage pools at a minimum: one primary, one on-site copy, and one off-site copy. I would actually like to have two off-site copy storage pools. Since this data doesn't change, and no additional backups will occur for the files, there will be no need for reclamation. The extra copy storage pools are a safety net in case we have bad media spots/tapes. Without reclamation, we will never know if we have bad media. So, at a minimum, 3 storage pools containing a total of 7.5+ million objects ((directory+files)*3) will use up 4.3GB of a TSM database (7.5 million * 600 bytes). The amount of growth per year is being estimated at about 4+ GB of TSM database, so, approximately another 2.3+ million files/folders each year. It will very likely be more. (Daily estimates are 6500 additional files/folders). Keep in mind. This data will NOT be changed or deleted in the foreseeable future. New data incoming daily. NO data expires. I don't know if Fuji will ever change the way they store their images/text files. So, here is what I am trying to figure out. 1. Will adding the additional objects from the PACS servers significantly increase my expiration processing run time? Will TSM have to scan all of those database objects during expiration processing? 2. I have heard it is possible to run another instance of TSM server on the same machine. Would that be a good idea here? It makes sense to this novice user. I wouldn't have to run expiration processing daily on the PACS TSM instance. 3. If a second TSM server instance is the recommended course, how difficult is that to setup? Any redbooks or how-tos out there regarding that? What issues do I have with sharing my LTO library between the two TSM server instances? Any redbooks or how-tos on sharing a single 3584 library (five LTO drives, and hoping to get more out of this project) between two TSM server instances on the same machine? 4. Regardless of how many TSM server instances, journaling will have to be setup on each of the NT4.0 PACS servers. What kind of overhead can we expect to run journaling on the NT servers (I haven't setup journaling anywhere, yet)? Three of the servers each have about 400-500K image objects, and the fourth server (the one with all of the 1k text files) has close to 1 million image/text objects (none of that includes the OS or application files/databases, just image/textreport files). 5. Due to the fact that the directory and file objects will likely not change, would there be a pressing need to use a DIRMC (non-migrating)? I would suppose not. 6. Is there a better way to do this? (probably should have asked that question first.) :+) Thanks in advance. Todd Lundstedt Technical Specialist Via Christi Information Management Services ofc. (316) 261-8385 fax. (316) 660-0036 [EMAIL PROTECTED]
TSM Upgrade frequency
For you folks that patch your TSM server frequently (like the individual from Oxford University who upgraded from some version to 5.1.6.0 on Feb3, then to 5.1.6.1 on Feb7, and now planning to upgrade to 5.1.6.2 within a couple weeks), how many of you have Storage Agent nodes? I am still on 4.2.1.7 simply because of the Storage Agent issue. We have a problem with TSM server crashes that is supposed to be fixed in 4.2.1.9+, but I find it hard to justify upgrading to 4.2.1.9+ when I know I will have to upgrade to 5.x soon after. According to the documentation I have read, the TSM Server version and the Storage Agent version must be at the same level. We have several Storage Agent nodes, and all of them are considered 24x7 servers, mission critical. It is like pulling teeth to get downtime for any one of them just for normal OS or Application type maint. I dread the time where I have to take them all down at once. Even more so, I dread having to take all of them down at the same time several times a month just to keep updating TSM Storage Agent versions because the TSM Server needed a patch. How do you folks with several mission critical, 24x7 servers go about doing TSM Server upgrades and TSM Storage Agent upgrades? Especially considering that TSM Server outages are best done during the day when no backups are occurring, and TSM Client Node outages are done during the night to reduce impact on the users of those Client Nodes.
Re: Another Expiration question
I am not certain on the order in which filespaces are processed during expiration, but, from looking at the log of my expiration processing, it appears to go in node (or perhaps filespace) order, based on when they were added to TSM (certainly not alphabetical). Oldest node (or filespace) gets processed first, newest gets processed last. I am unaware if there is a flag indicating where you would like to start. I doubt there is, though. If you never let the expiration processing complete, you will never expire all of the files that can be expired. Hope this helps. |+ || Michael Raine | || Michael.Raine@ALL| || FIRST.COM| || Sent by: ADSM: | || Dist Stor Manager| || [EMAIL PROTECTED]| || EDU | ||| ||| || 02/14/2003 11:40 | || AM| || Please respond to | || ADSM: Dist Stor | || Manager | ||| |+ -| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Another Expiration question | -| Does anyone know the particulars on how expiration works. Normally we are run expiration with a time limit or 6 hours. When we ran it without the time limit it ran for 15 hours and never finished. Does expiration start reviewing the db from the beginning or is there some sort of flag set on where to start. I'm trying to figure out if we never let it run through completely will I ever get back all the data that has expired or not. Thanks for the help
Re: management class question???
Or, just back it up to diskpool. When the diskpool gets to the high water mark, it will start migration to the next storage pool (I assume, your tapepool). When the diskpool gets completely full, TSM will automatically send files to the next storage pool (again, assuming your tapepool), meaning, directly to tape. This is all good, as long as you have at least two tape drives. One thing I would suggest is to run this backup outside of your normal scheduled backups, so you will have this done (and hopefully copied to your copy stgpool (offsite)) before the main backups for all your other clients begin. Hope this helps. Todd Ford, Phillip phillip.ford@S To: [EMAIL PROTECTED] PCORP.COMcc: Sent by: ADSM: Fax to: Dist Stor Subject: Re: management class question??? Manager [EMAIL PROTECTED] ST.EDU 01/31/2003 10:41 AM Please respond to ADSM: Dist Stor Manager Just change the destination pool of the given management class to the tape pool (the one that the disk pool will migrate to). Make sure you activate the policy set. Then do your first backup. Now go in and change the destination pool to the disk pool. Make sure to activate the policy set again. That should do it. -- Phillip Ford Senior Software Specialist Corporate Computer Center Schering-Plough Corp. (901) 320-4462 (901) 320-4856 FAX [EMAIL PROTECTED] -Original Message- From: Joni Moyer [mailto:[EMAIL PROTECTED]] Sent: Friday, January 31, 2003 10:19 AM To: [EMAIL PROTECTED] Subject: management class question??? Hi everyone! I am creating a new management class for a user that wants all files backed up incrementally no matter if it is open or not. I know that I have to use the dynamic parameter for copy serialization. On a normal day I want the backup to go to the disk pool, but the first time this server backs up it will back up the entire 202 GB server. My problem is that the disk pool is only 34 GB. I want to direct the first backup directly to tape and all future backups to disk. How do I do this? If I tell the user to first use a MC that goes directly to tape and then use a MC that goes to disk for all others, will it not try to do a backup of all files for the second incremental backup since it is technically using a new MC class or does it go by files that exist? Thanks in advance for any help!!! Joni Moyer Systems Programmer [EMAIL PROTECTED] (717)975-8338 * This message and any attachments is solely for the intended recipient. If you are not the intended recipient, disclosure, copying, use or distribution of the information included in this message is prohibited -- Please immediately and permanently delete.
Re: AIX question
Becky (and Justin from another post), Both of you suggest that I first delete the offending disks... yet that is my exact problem. I can't delete them. When I try to rmdev them, I get an error indicating that there is not enough information to uniquely identify the device (rmdev -l 'hdisk2' -d ... but there are two hdisk2 devices). I was looking for a way to delete them other than rmdev. The AIX guru in our shop finally came up with something he remembered from his classes... odmget and odmdelete commands. They appear to allow one to more uniquely identify a single object, more so than rmdev. Are there any gotchas to consider while using odmget and odmdelete? Todd Davidson, BeckyTo: [EMAIL PROTECTED] Becky.Davidson cc: @SLBG.COMFax to: Sent by: ADSM: Subject: Re: AIX question Dist Stor Manager [EMAIL PROTECTED] ST.EDU 01/06/2003 06:47 PM Please respond to ADSM: Dist Stor Manager Delete the two hdisk again. Run cfgmgr again. Check it and if you see the same type of pattern of 1 vpath showing 4 and the other 2 showing only one then the shark guy needs to recheck his configuration. If you see a different pattern try running cfgmgr again. There was a problem awhile back that you need to run cfgmgr x's with x being the number of paths so that you would get all of the paths. Sometimes we have had to delete and run cfgmgr again because it doesn't seem to work right the first time. After delete and rerunning cfgmgr it seems to work better. Something you might want to get your hands on is the Shark Gotcha's document. I think it is still on the storage page but if not email me off list and I will send it to you. Another issue you might run into is that the devices may not be fully formatted. There was a version of the shark code that said that everything was formatted yet it wasn't. The only way you could tell was to go into the shark under where you configure paths and refresh the screen. good luck and let me know if I can be of further assistance Becky -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED]] Sent: Monday, January 06, 2003 4:50 PM To: [EMAIL PROTECTED] Subject: AIX question First, let me say.. Dear Lord, the IBM web pages are a pain to dig through and find what you are looking for. Couldn't find any knowledge bases or anything to help me resolve the issue I have... so.. here goes. AIX 4.3.3 My TSM server is a pSeries with two Fibre Channel Adapters installed, through which we access data on the ESS (shark) and the LTO tape drives (I know, bad idea). Each of the fibre cards is connected to its own fibre switch. Each switch has two connections to the Shark (four data paths to the Shark). We have been trying to add some additional Shark storage to the TSM server. After the Shark guy did his thing, I ran cfgmgr, and thinks didn't look quite right. The two old vpath devices each showed four disks (hdsikX), two on each of the two fibre cards (fscsi1, fscsi0), just like they always have. One of the three new vpath devices showed four hdisks on two different fscsi devices, but the other two vpath devices only showed two hdisks, one on each fscsi. lscfg showed a bunch of new disks, mostly available, but four were defined. I decided to delete the three new vpath devices, and all of the associated hdisk devices, including the defined ones. As I got to looking, I saw something odd... root@ax-dco-tsm / # lsdev -Cc disk hdisk0 Available 10-88-00-8,0 16 Bit LVD SCSI Disk Drive hdisk1 Available 10-88-00-9,0 16 Bit LVD SCSI Disk Drive hdisk18 Available 10-68-01 IBM FC 2105F20 hdisk19 Available 10-68-01 IBM FC 2105F20 hdisk20 Available 10-68-01 IBM FC 2105F20 hdisk21 Available 10-68-01 IBM FC 2105F20 hdisk22 Available 10-78-01 IBM FC 2105F20 hdisk23 Available 10-78-01 IBM FC 2105F20 hdisk24 Available 10-78-01 IBM FC 2105F20 hdisk25 Available 10-78-01 IBM FC 2105F20 vpath3 Available Data Path Optimizer Pseudo Device Driver vpath4 Available Data Path Optimizer Pseudo Device Driver hdisk2 Defined 10-78-01 IBM FC 2105F20 hdisk2 Defined 10-68-01 IBM FC 2105F20 hdisk3 Defined 10-68-01 IBM FC 2105F20 hdisk3 Defined 10-68-01 IBM FC 2105F20 root@ax-dco-tsm / # datapath query device Total Devices : 2 DEV#: 0 DEVICE NAME: vpath3 TYPE: 2105F20 SERIAL: 00719516 Path# Adapter/Hard DiskState Mode Select Errors 0 fscsi0/hdisk18 OPEN NORMAL 25 0 1 fscsi0/hdisk20 OPEN
AIX question
First, let me say.. Dear Lord, the IBM web pages are a pain to dig through and find what you are looking for. Couldn't find any knowledge bases or anything to help me resolve the issue I have... so.. here goes. AIX 4.3.3 My TSM server is a pSeries with two Fibre Channel Adapters installed, through which we access data on the ESS (shark) and the LTO tape drives (I know, bad idea). Each of the fibre cards is connected to its own fibre switch. Each switch has two connections to the Shark (four data paths to the Shark). We have been trying to add some additional Shark storage to the TSM server. After the Shark guy did his thing, I ran cfgmgr, and thinks didn't look quite right. The two old vpath devices each showed four disks (hdsikX), two on each of the two fibre cards (fscsi1, fscsi0), just like they always have. One of the three new vpath devices showed four hdisks on two different fscsi devices, but the other two vpath devices only showed two hdisks, one on each fscsi. lscfg showed a bunch of new disks, mostly available, but four were defined. I decided to delete the three new vpath devices, and all of the associated hdisk devices, including the defined ones. As I got to looking, I saw something odd... root@ax-dco-tsm / # lsdev -Cc disk hdisk0 Available 10-88-00-8,0 16 Bit LVD SCSI Disk Drive hdisk1 Available 10-88-00-9,0 16 Bit LVD SCSI Disk Drive hdisk18 Available 10-68-01 IBM FC 2105F20 hdisk19 Available 10-68-01 IBM FC 2105F20 hdisk20 Available 10-68-01 IBM FC 2105F20 hdisk21 Available 10-68-01 IBM FC 2105F20 hdisk22 Available 10-78-01 IBM FC 2105F20 hdisk23 Available 10-78-01 IBM FC 2105F20 hdisk24 Available 10-78-01 IBM FC 2105F20 hdisk25 Available 10-78-01 IBM FC 2105F20 vpath3 Available Data Path Optimizer Pseudo Device Driver vpath4 Available Data Path Optimizer Pseudo Device Driver hdisk2 Defined 10-78-01 IBM FC 2105F20 hdisk2 Defined 10-68-01 IBM FC 2105F20 hdisk3 Defined 10-68-01 IBM FC 2105F20 hdisk3 Defined 10-68-01 IBM FC 2105F20 root@ax-dco-tsm / # datapath query device Total Devices : 2 DEV#: 0 DEVICE NAME: vpath3 TYPE: 2105F20 SERIAL: 00719516 Path# Adapter/Hard DiskState Mode Select Errors 0 fscsi0/hdisk18 OPEN NORMAL 25 0 1 fscsi0/hdisk20 OPEN NORMAL 31 0 2 fscsi1/hdisk22 OPEN NORMAL 29 0 3 fscsi1/hdisk24 OPEN NORMAL 26 0 DEV#: 1 DEVICE NAME: vpath4 TYPE: 2105F20 SERIAL: 60219516 Path# Adapter/Hard DiskState Mode Select Errors 0 fscsi0/hdisk19 OPEN NORMAL 13465 0 1 fscsi0/hdisk21 OPEN NORMAL 13636 0 2 fscsi1/hdisk23 OPEN NORMAL 13567 0 3 fscsi1/hdisk25 OPEN NORMAL 13727 0 ++ Note the lines. I have two entries for hdisk2 and two for hdisk3. Smitty won't let me remove them (not enough information to uniquely identify the device) and similar error messages for change/show charactaristics. How can I remove hdisk2 and hdisk3? The 10-78 and 10-68 devices are the fibre cards. Thanks in advance Todd Lundstedt Technical Specialist Via Christi Information Management Services ofc. (316) 261-8385 fax. (316) 660-0036 [EMAIL PROTECTED]
Mulitple Schedule backup strategy
I am trying to see if I am on the right track here. I don't have a test Netware server to play with. Please let me know if I am thinking correctly, or if I am way off base. Thanks I have recently been told that we need to backup certain files on a machine weekly, and the rest of the files on the same machine daily. I think the way to go about this is to add another nodename in the dsm.opt file (Netware server here), along with the excludes/includes for that node. So, the dsm.opt my NW-DCO-Server4 would look something like this... +++ * Tivoli Storage Manager COMMMETHODTCPip TCPSERVERADDRESS 200.1.1.41 TCPPORT 1500 NODENAME NW-DCO-Server4 EXCLUDE sys:\system\secaudit.log EXCLUDE sys:\system\events.log EXCLUDE sys:\system\system.log EXCLUDE sys:\system\btrieve.trn EXCLUDE SYS:\system\tsa\err$hst.* EXCLUDE sys:\system\tsa\err$log.* EXCLUDE sys:\system\tsa\skip$log.* EXCLUDE sys:\system\tsa\tsa$temp.* EXCLUDE sys:\system\sys$log.err EXCLUDE sys:\_swap_.mem EXCLUDE sys:\vol$log.err EXCLUDE sys:\tts$log.err EXCLUDE VOL1:\DEPT\VOLNTEER\VWORKS\...\* NWPWFILE YES QUIET PASSWORDACCESS GENERATE SCHEDLOGRETENTION 5 NWIGNORECOMPRESSBIT YES NODENAME NW-DCO-Server4.a EXCLUDE * INCLUDE VOL1:\DEPT\VOLNTEER\VWORKS\...\* NWPWFILE YES QUIET PASSWORDACCESS GENERATE SCHEDLOGRETENTION 5 NWIGNORECOMPRESSBIT YES +++ I would have to define the node NW-DCO-Server4.a to TSM, and schedule it for weekly backups with the original schedule for daily backups. Am I on the right track here? Am I missing something? Is there a different/better way to do this? Thanks Todd
Re: Idle system fails with Media mount not possible
How strange.. I just went through something similar. Running on AIX 4.3.3, TSM 4.2.1.7. What are you running? The short answer was to reboot the AIX operating system, and everything started working fine.. The long answer follows (well, not really an answer, just my situation, and what I tried to resolve it). Server AIX 4.3.3 TSM 4.2.1.7 Nodes W2K Storage Agent 4.2.1.7 BA Client 4.2.1.32 TDP for SQL 2.2 SQL 2000 and WinNT4 Storage Agent 4.2.1.7 BA Client 4.2.1.15 TDP for SQL 1.1 SQL 6.5 Relevant TSM server storage as follows... diskpool_sql_meta (no next storage pool intended only for the */.../meta/.../* info) diskpool_sql (next storage pool is ltotape_sql, intended for smaller databases) ltotape_sql (collocation of FILESPACE since /stripes=2 backups are kept here) The SQL 2000 server had been having issues over the last few months where backups to ltotape_sql with /stripes=2 of a 265GB database would fail with a server media mount no possible error, but /stripes=1 differential backups would back up fine. Oddly, increasing the Maximum Mount Points for the node by one would allow the /stripes=2 backup to succeed, but the next time a /stripes=2 backup would run, it would fail (until I increased the MMP again). I had 5 drives, all free and unused and 7 MMP for the node when... this new wrinkle occurred. The SQL 6.5 server started having problems backing up certain databases: the smaller system databases; master, model, msdb, pubs, tempdb, with and error message of server media mount not possible. All the DBs on this server have a destination of ltotape_sql. Like you, plenty of room in the storage pool, plenty of scratch. Called support Got level one.. told him a few things.. he didn't even want to try it.. and immediately escalated to level two. While I waited for a call back from level two, the following occurred. I noticed that there are some databases in diskpool_sql that haven't migrated to ltotape_sql. Kicking off a migration gets a similar error message media mount not possible, which, oddly, is the same message I got from the storage agent when backing up tapes to ltotape_sql. I carefully detailed what it took to migrate those 3 files from diskpool_sql to ltotape_sql, which is a whole other chapter by itself, involving changing maxscratch up and down, moving data, and a few other hoops, and I was unable to get some tapes to move with a move data command (tapes that had only one master or msdb or tempdb type database on them). Level two calls back. I go through the entire situation, including the fact of the Max Mount Point having to change every time I did I /stirpes=2 backup (I wasn't sure if that was a related issue or not). She is baffled, and wants to think it over and search databases etc to see what she can come up with. Within 30 mins, she calls back and asks me to reboot the TSM server's OS (uptime reported a whopping 82 days), just to see what would happen. I do. Migrations go. Backups /stripes=1 go. Backups /stripes=2 go (even with MMP set back to 4 for that node, instead of 7 ( with only 5 tape drives remember). This was Friday. Sunday night, the TSM server did something odd (haven't reported this to TSM support yet). It just stopped. It showed link status on the fiber cards, and network cards, but you couldn't ping it, the server console wouldn't wake up, nothing. Even the display on the front was dark, but the power light was on steady like it was operational, not flashing like it would be if you did a proper shutdown. I reset it Monday morning when I found it that way, and then had to do a clean shutdown and power on to get the fiber cards to see the library correctly. Very weird. So, I am taking Monday morning (yesterday) as the start time to see how long it takes until I have to increase my MMP on the one node just to get a /stripes=2 backup. The saga continues... Conko, Steven To: [EMAIL PROTECTED] [EMAIL PROTECTED] cc: M Fax to: Sent by: Subject: Idle system fails with Media mount not possible ADSM: Dist Stor Manager [EMAIL PROTECTED] IST.EDU 12/17/2002 03:19 PM Please respond to ADSM: Dist Stor Manager strange one... and ive looked at everything i can think of. In client dsmerror.log: 12/17/02 15:01:54 ANS1228E Sending of object '/tibco/logs/hawk/log/Hawk4.log' failed 12/17/02 15:01:54 ANS1312E Server media mount not possible 12/17/02 15:01:57 ANS1312E Server media mount not possible In activity log: ANR0535W Transaction failed for session 1356 for node SY00113 (AIX) - insufficient mount points available to satisfy the request. There is NOTHING else running on this TSM server. All 6
Re: include/exclude syntax question for TSM Unix Client
I believe exclude.dir is processed first regardless of where it is stacked in the list. You will have to work your exclude.dir statements so there isn't one that will hit true on the u38 directory; exclude.dir /u[0-2,4-9][0-9] exclude.dir /u3[0-7,9] Hope this helps. Peppers, Holly Holly.Peppers@B To: [EMAIL PROTECTED] CBSFL.COM cc: Sent by: ADSM:Fax to: Dist Stor Subject: include/exclude syntax question for TSM Unix Client Manager [EMAIL PROTECTED] T.EDU 12/04/2002 03:52 PM Please respond to ADSM: Dist Stor Manager Hi All, I was hoping I could get some feedback on the syntax of a set of files that I am trying to include during backup. Currently, my exclude.list file looks like so: exclude.dir /u[0-9][0-9] include.file /u38/exp/.../*.* I'm trying to get TSM to backup all the files and subdirs under /u38/exp, but still exclude all the the other u* directories and files. I must be doing something wrong, as TSM is still excluding /u38/exp. I've tried a few different combinations, but with no luck. Does anyone have any suggestions? Thanks. Holly L. Peppers BCBSFL Cpacity Planning Blue Cross Blue Shield of Florida, Inc., and its subsidiary and affiliate companies are not responsible for errors or omissions in this e-mail message. Any personal comments made in this e-mail do not reflect the views of Blue Cross Blue Shield of Florida, Inc.
Re: System Object / Mgmt. Classes / Policy Sets / Backup Groups - - Wh at's the best way??
Correct me if I am wrong, but this will only change the management class for the backups after the changes are made. All of the other systemobjects that exist in the database that are managed by the old management class will be retained for their 60 days. In order to facilitate the move to the new management class, you will want to rename the systemobject filespace on each server (for instance, systemobject.old), then perform your backups for 8-9 days, making sure the new systemobject only has data stored for 7 days, and only in the new management class. Then, it will be safe to delete systemobject.old, and recover that database space prior to your upgrade. Todd |+-- || Prather, Wanda| || Wanda.Prather@J| || HUAPL.EDU | || Sent by: ADSM: | || Dist Stor | || Manager| || [EMAIL PROTECTED]| || T.EDU | || | || | || 10/23/2002 11:48| || AM | || Please respond | || to ADSM: Dist | || Stor Manager | || | |+-- | | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: System Object / Mgmt. Classes / Policy Sets / Backup Groups - - Wh at's the best way?? | | Hi Kevin, Most people have only 2 policy sets for each policy domain - the acitve and the inactive one. I suggest you make a new management class: * In the NOCODOM policy domain, create a new management class, for example call it SYSOBJ_MAN * Edit the BACKUP COPY GROUP for that new management class to set your retention to 7 instead of 60. * ACTIVATE the policy set for that domain; that makes the changed version active. * Pick a Win2K client in the NOCODOM domain to test. * In the dsm.opt file of that client, put an include statement to cause TSM to bind the systemobject to the new management class: include.systemobject SYSOBJ_MAN * Start the backup client and backup the system object * Now pretend you are going to restore the system object: start the client click RESTORE expand SYSTEM OBJECT click on one of the components (like registry) scroll to the right, you should see the time and management class of the last backup if it says SYSOBJ_MAN instead of DEFAULT, it's working * Now you have to figure out how you are going to make this happen on ALL the clients in the NOCODOM domain. * You can either use sneakernet and add this to the dsm.opt file of all the affected clients, or create a clientoptionset on the server with this include statement (certainly easier to do!) -Original Message- From: Thach, Kevin [mailto:KThach;COVHLTH.COM] Sent: Monday, October 21, 2002 3:54 PM To: [EMAIL PROTECTED] Subject: System Object / Mgmt. Classes / Policy Sets / Backup Groups -- Wh at's the best way?? Hi, I've been designated the TSM administrator here at my workplace, and I'm trying to educate myself as I go without making a mess of things. Our IBM partner came in and installed the server and a few clients, and handed it over to me. I now have about 200 Win2K clients, 40 AIX Clients, and about 5 Linux Clients. My question is this: The person that installed our environment basically set up 6 Policy Domains: Colodom, Exchange, Lanfree, MSSQL, Nocodom, and Oracle. 99% of the clients are in the Nocodom (non-collocated) domain, which has one policy set, and one management class which has one backup copy group with retention policies set to NOLIMIT, 3, 60, 60. So, my problem is this. It turns out that in trying to upgrade from 4.2.2.3 to 5.1.X we ran into the problem with System Objects BIGTIME. We are retaining 60 days worth of System Object files for about 200 Win2K clients, which translates to about 29,000,000 System Object files, and about 55% of our TSM database (36 GB). What we've decided to do before
Re: OK IBM _ ENOUGH ALREADY with the IBM/Tivoli TSM website -
I think that may be because all of the functionality at Tivoli.com has been (or soon will be completely) replaced with functionality at IBM.com. I received several emails over the last month regarding this last week, and I think Friday, Oct. 18 was officially the last day we were to try to use Tivoli.com site. Just my understanding of the emails I have been receiving... Todd |+-- || Prather, Wanda| || Wanda.Prather@J| || HUAPL.EDU | || Sent by: ADSM: | || Dist Stor | || Manager| || [EMAIL PROTECTED]| || T.EDU | || | || | || 10/23/2002 02:42| || PM | || Please respond | || to ADSM: Dist | || Stor Manager | || | |+-- | | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: OK IBM _ ENOUGH ALREADY with the IBM/Tivoli TSM website - | | I don't have any trouble going to the FTP site, THANK GOODNESS!! I'm talking about the Tivoli web pages on the IBM support site. A lot of new users are not ready to navigate the FTP site and figure out patch/fix levels on their own; we have taught many support people to start with the Tivoli web pages, which have NICE select drop downs to help them figure out the current levels of code they need; when they click the web page, it sends them to the RIGHT PLACE on the FTP site. Either the web pages should WORK, or they should be eliminated entirely. They should not be implemented in a broken state. This is the web site I'm talking about: http://www.tivoli.com/support/storage_mgr/adsercli.htm Start here, select the WinNT-2000 , then try going through the SUPPORT DOWNLOADS link -Original Message- From: Richard Cowen [mailto:richard_cowen;CNT.COM] Sent: Wednesday, October 23, 2002 3:15 PM To: [EMAIL PROTECTED] Subject: Re: OK IBM _ ENOUGH ALREADY with the IBM/Tivoli TSM website - I downloaded the AIX server v5151 yesterday, I can navigate the site today. ftp index.storsys.ibm.com Connected to serviceb.boulder.ibm.com. 220 serviceb.boulder.ibm.com FTP server (Version wu-2.6.2(1) Mon Dec 3 15:26:19 MST 2001) ready User (serviceb.boulder.ibm.com:(none)): ftp 331 Guest login ok, send your complete e-mail address as password. Password: 230 Guest login ok, access restrictions apply. ftp cd tivoli-storage-management 250 CWD command successful. ftp dir 200 PORT command successful. 150 Opening ASCII mode data connection for /bin/ls. total 88 -rw-rw-r-- 1 18125700 200 490 Jun 28 2001 .message -rw-r-x--- 1 0200 71 Jan 24 2000 .profile d--x--s--x 2 0200 512 Jan 24 2000 bin d--x--s--x 2 0200 512 Jan 24 2000 dev drwxrwsr-x 3 18125700 200 512 Jun 26 2001 devices d--x--s--x 2 0200 512 Jan 24 2000 lib drwxrwsr-x 9 18125700 200 512 Jul 31 2001 maintenance drwxrwsr-x 9 18125700 200 512 Apr 5 2002 patches drwxrwsr-x 2 18125700 200 512 Nov 29 2000 plus drwxrwsr-x 2 18125700 200 8192 Oct 22 17:05 swap 226 Transfer complete. ftp: 628 bytes received in 0.00Seconds 628000.00Kbytes/sec. ftp cd patches 250 CWD command successful. ftp cd client 250 CWD command successful. ftp dir 200 PORT command successful. 150 Opening ASCII mode data connection for /bin/ls. total 32 drwxrwsr-x 10 18125700 200 512 Mar 30 2001 v3r7 drwxrwsr-x 12 18125700 200 512 Jun 11 10:49 v4r1 drwxrwsr-x 13 217 200 512 May 31 06:05 v4r2 drwxrwsr-x 9 18125700 200 512 Oct 18 11:03 v5r1 226 Transfer complete. ftp: 250 bytes received in 0.01Seconds 25.00Kbytes/sec. ftp cd v5r1 250 CWD command successful. ftp dir 200 PORT command successful. 150 Opening ASCII mode data connection for /bin/ls. total 80832 -rw-rw-r-- 1 18125700 200 41166639 Oct 17 17:04 IP22546_02.exe -rw-rw-r-- 1 18125700 20091223 Oct 17 17:04
Re: Audit Library question.
The reason I started the audit was because TSM was not reporting the tape in the library, yet the library knew the tape was inserted. I could see the tape in the library (with my own eyes). Using the manual operations | move tape functions from the LCD display on the library, the library was able to move the tape out and back into the library. But Query LIBVolume did not show the tape in the library. I thought Audit Library with checklabel=barcode should fix it, but after 2 hours, the process hadn't ended. So I cancelled it. What I ended up doing was manually removed the tape (via the move tape functions from the LCD panel of the library), and then turned around and did a Checkin process for the tapes in the Bulk I/O slots. After that, the query libvolume command reported the tape in the library. This tells me that there is/was no problem with the barcode, or the reader, and possibly even the library memory (since the library knew it had the tape all along). Something funky going on with the Audit Library process, for sure. |+-- || David Longo | || David.Longo@HEALTH-| || FIRST.ORG | || Sent by: ADSM: Dist| || Stor Manager | || [EMAIL PROTECTED]| || U | || | || | || 10/22/2002 01:02 PM | || Please respond to | || ADSM: Dist Stor| || Manager| || | |+-- ---| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: Audit Library question. | ---| I imagine the checkl=barocde was introduced to shorten audit, without it you would have to mount every tape in library - which would take some considerable time with some libraries! What you are doing is checkinbg the barcode label in library memory as opposed to checking the magnetic tape label header. The ideal short way is to have the library do it's inventory, which reads barcodes and is quick, then do audit with checkl=barcode. Whole process shouldn't take more than a few minutes - there may be some library units that take longer. This complete process should take care of anything that has gotten out of sync. I have had a few cases where there was still something out of sync and had to do detailed examination to correct. It can have a problem reading the barcode if the laser scanner couldn't read the label. That can happen some times - especially if you don't use original manufacturers labels. If you have AIX server and use tapeutil with inventory action, it will show the slot status for tapes like these in abnormal status. When the audit with checkl=barcode runs it finds this and no barcode label for that slot and mounts the tape in that slot to read the magnetic label and update TSM's inventory. A brief overview as I have seen it in action many times. David B. Longo System Administrator Health First, Inc. 3300 Fiske Blvd. Rockledge, FL 32955-4305 PH 321.434.5536 Pager 321.634.8230 Fax:321.434.5509 [EMAIL PROTECTED] [EMAIL PROTECTED] 10/22/02 01:44PM At 11:29 AM -0400 10/22/02, David Longo said: With checklabel=barcode, what happens is that TSM reads the internal memory of the library as to what the library's inventory says is where. So checklabel=barcode doesn't really mean read the barcodes? It just means check the library's internal memory? I guess that's still useful in some circumstances, if there'e a possibility that TSM and the library have gotten out of sync. But it would be nice if things mean what they say. Suppose I really want it to read the barcodes? Suppose I think the library's internal memory has gotten confused somehow, and I want to do a physical audit of barcode locations to compare with the internal memory? Is this possible? Or is it a function of the library (which I guess might make more sense). So generally that won't take long. And a drive needs to be available for the case where library had a problem
Re: Audit Library question.
Not true on my library, Ken... I have run several audits using checklabel=barcode before with success. The arm has never moved in the library with an audit using checkl=b. |+ || KEN HORACEK | || KHORACEK@INCS| || YSTEM.COM| || Sent by: | || ADSM: Dist | || Stor Manager | || [EMAIL PROTECTED]| || IST.EDU | ||| ||| || 10/22/2002| || 01:03 PM | || Please respond| || to ADSM: Dist| || Stor Manager | ||| |+ ---| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: Audit Library question. | ---| Not true... With checklabel=barcode, all of the barcodes are read. This is then checked with the internal memory of the library as to what the library's inventory says is where. The tape is mounted, only if the barcode is mis-read. Ken [EMAIL PROTECTED] [EMAIL PROTECTED] 10/22/2002 10:44:50 AM At 11:29 AM -0400 10/22/02, David Longo said: With checklabel=barcode, what happens is that TSM reads the internal memory of the library as to what the library's inventory says is where. So checklabel=barcode doesn't really mean read the barcodes? It just means check the library's internal memory? I guess that's still useful in some circumstances, if there'e a possibility that TSM and the library have gotten out of sync. But it would be nice if things mean what they say. Suppose I really want it to read the barcodes? Suppose I think the library's internal memory has gotten confused somehow, and I want to do a physical audit of barcode locations to compare with the internal memory? Is this possible? Or is it a function of the library (which I guess might make more sense). So generally that won't take long. And a drive needs to be available for the case where library had a problem reading a barcode label, that tape can be mounted in a tape drive to verify - even if using checkl=b. But how can it have a problem reading the barcode label if check-=b doesn't even try to read the labels? -- Matt Simpson -- OS/390 Support 219 McVey Hall -- (859) 257-2900 x300 University Of Kentucky, Lexington, KY 40506 mailto:msimpson;uky.edu mainframe -- An obsolete device still used by thousands of obsolete companies serving billions of obsolete customers and making huge obsolete profits for their obsolete shareholders. And this year's run twice as fast as last year's. - This e-mail, including attachments, may include confidential and/or proprietary information, and may be used only by the person or entity to which it is addressed. If the reader of this e-mail is not the intended recipient or his or her authorized agent, the reader is hereby notified that any dissemination, distribution or copying of this e-mail is prohibited. If you have received this e-mail in error, please notify the sender by replying to this message and delete this e-mail immediately. - GWIASIG 0.07
Re: offsite tape destroyed
When dealing with multiple primary storage pools being copied to a single copy storage pool, it might be better to use move data. And yes, you can move data from a copy storage pool volume as long as it is set to access=offsite (unavailable works, too, I think). move data will work OR delete volume x discarddata=yes then backup stg primary_stgpool copy_stgpool (for EACH primary_stgpool) will work. |+-- || David Longo | || David.Longo@HEALTH-| || FIRST.ORG | || Sent by: ADSM: Dist| || Stor Manager | || [EMAIL PROTECTED]| || U | || | || | || 10/04/02 10:20 AM | || Please respond to | || ADSM: Dist Stor| || Manager| || | |+-- ---| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: offsite tape destroyed | ---| If an offsite tape (COPYPOOL) is destroyed then it is simple. Do a DEL VOL blah discarddata=yes, for the destroyed volume. When that is complete then do a Backup STG blah blah and this will recopy the data (and any data not copied) from the onsite pool to a new tape on the offsite pool. The TSM DB keeps track and knows that the copy has been deleted so it recopies at next backup stg process run. You can't do a Move Data because the original volume is destroyed. David Longo [EMAIL PROTECTED] 10/04/02 10:54AM I did a quick search of some stuff I have and can't find an answer, plus I just don't have time to sit and look this morning so I need a bit of help. If an offsite tape is destroyed what are the steps to deal with this. I know it's not the same for a primary stg pool. Thanks, Geoff Gill TSM Administrator NT Systems Support Engineer SAIC E-Mail: [EMAIL PROTECTED] Phone: (858) 826-4062 Pager: (877) 905-7154 MMS health-first.org made the following annotations on 10/04/2002 11:21:54 AM -- This message is for the named person's use only. It may contain confidential, proprietary, or legally privileged information. No confidentiality or privilege is waived or lost by any mistransmission. If you receive this message in error, please immediately delete it and all copies of it from your system, destroy any hard copies of it, and notify the sender. You must not, directly or indirectly, use, disclose, distribute, print, or copy any part of this message if you are not the intended recipient. Health First reserves the right to monitor all e-mail communications through its networks. Any views or opinions expressed in this message are solely those of the individual sender, except (1) where the message states such views or opinions are on behalf of a particular entity; and (2) the sender is authorized by the entity to give such views or opinions. ==
Re: Anyone tried 5.1.1.6?
I am interested in reading your write-up. Is it complete, and if so, under what subject? If not, do you have an estimate on when? Thanks for doing this Gretchen, Todd |+ || Gretchen L. | || Thiele | || GRETCHEN@PUCC| || .BITNET | || Sent by: | || ADSM: Dist | || Stor Manager | || [EMAIL PROTECTED]| || IST.EDU | ||| ||| || 09/24/02 08:47| || PM| || Please respond| || to ADSM: Dist| || Stor Manager | ||| |+ -| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Re: Anyone tried 5.1.1.6? | -| I found that the SYSTEM OBJECT which were not unicode were never deleted unless I did it manually. A QUERY OCC * SYSTEM OBJECT (without the NAMETYPE=UNICODE option) will let you know if you have that problem. Also, did you run the command CLEANUP BACKUPGROUPS? For me, on 5.1.1.4 it ran but never cleaned everything up and it wouldn't run on 5.1.1.6. I finally had to do an AUDITDB FIX=YES. Actually, this problem occurs with both unicode and nonunicode SYSTEM OBJECT filesystems. If you delete the nonunicode filesystem it will rebackup as unicode. The new unicode SYSTEM OBJECT filesystem will continue accumulate extra versions as well. I could run CLEANUP BACKUPGROUPS with an earlier v5 version, but it didn't work (I ran it, but then ran it again and it found more errors). The CLEANUP didn't work (got the loop problem) under v5.1.1.6, but then the AUDITDB didn't work! This is what I'm writing up... Gretchen Thiele Princeton University
Re: Is reclamation working?
I have seen the number of reclaimable tapes grow steadily over the course of a week, but the percent utilized average would go down over time. If I remember correctly, reclaiming tapes in offsite pools is similar to a migration, in that the reclamation process moves the data in the largest chunks possible. For example... Day one, you have 3 tapes that need to be reclaimed in the offsite pool, but by the end of the day, they each only get down to 25-30% utilized. Day two, three more tapes are added to the reclamation list, but since onsite tapes mounted for reclamation will have more data to be moved on them for the newly added tapes, the tapes from day one will not get reclaimed as much as the tapes from day two. So now you end up with the day one tapes an additional 5% less utilized, and the day two tapes 20% less utilized. Day three comes along, adding four tapes that are 50% utilized, and the list of tapes grows, but the reclaimable percentage, on average, is still less. Each day, this happens without fully reclaiming some tapes. What I do when I see this condition is set the reclamation threshold higher.. say 95% or 90%.. and get the tapes that are 5 or 10% utilized fully reclaimed. Later, I will set the threshold lower, in increments that make sense (say, only two or four tapes will match the reclamation threshold) so that tapes get fully reclaimed. The situation you are seeing can happen very easily if your primary storage pool is collocated, but your copy storage pool is not. Lots of primary storage pool tape mounts will slow down your reclamation process. Hope this helps Todd |+-- || Gill, Geoffrey | || L. | || GEOFFREY.L.GILL| || @SAIC.COM | || Sent by: ADSM: | || Dist Stor | || Manager| || [EMAIL PROTECTED]| || T.EDU | || | || | || 09/11/02 10:05 | || AM | || Please respond | || to ADSM: Dist | || Stor Manager | || | |+-- -| | | | To: [EMAIL PROTECTED] | | cc: | | Fax to: | | Subject: Is reclamation working? | -| TSM Server is AIX 4.3.3 running TSM 4.2.1.9 I have noticed I have about 145 volumes that are less than 25% utilized, most of these are in the offsite pool. Reclamation runs on the offsite pool daily, sometimes, like this morning, I find it is still running. Reclamation kicks off at 5PM with reclaim=50. I'm not understanding why these aren't being reclaimed quicker. Just a few: U00582OFFSITE_POOL 3590CLASS25,415.90.0 Full U00586OFFSITE_POOL 3590CLASS19,665.60.0 Full U01638OFFSITE_POOL 3590CLASS18,157.98.5 Full This morning I saw reclamation running with this tape mounted and a scratch tape had just been mounted to, more than likely, move some files. But it's 100% full so I don't get it. U02188NOTES_TPOOL 3590CLASS36,672.3 100.0Full Is this strange and does anyone else have this many tapes with the same percentage used. Geoff Gill TSM Administrator NT Systems Support Engineer SAIC E-Mail:mailto:[EMAIL PROTECTED] [EMAIL PROTECTED] Phone: (858) 826-4062 Pager: (877) 905-7154
Backups of a file when it hasn't changed.
Does anyone else see this issue? I have seen several instances where TSM backed up a file when it had not been modified in anyway. Even when you look at the restore application, there are several versions of a file, each with the same create date, the same modified date, and different backed up dates. I am curious if others, or everyone sees this happening, or if I have some setting incorrect some where. There have been no always backup backups (selective) performed, so that isn't what happened. Clueless... Todd
Re: Backups of a file when it hasn't changed.
I had considered that, and honestly, I wouldn't know how to tell if the permissions had changed on the days in question. I don't see why they would have changed. There are other files in the same directory that have their only backups on the same day as the oldest backups for these files in question (the day this node started backing up to TSM). Yet some of the files have had additional backups when it doesn't appear to have been necessary. We don't normally set permissions at the file level (this is a Netware server). Plus, some of the files in question wouldn't even have a need to be accessed.. things like Ghost(R) files and old archival type documents. It really boggles the mind. Todd |+- || Ford, Phillip| || phillip.ford@S| || PCORP.COM | || Sent by: ADSM:| || Dist Stor | || Manager | || [EMAIL PROTECTED]| || ST.EDU| || | || | || 08/02/02 02:53 | || PM | || Please respond | || to ADSM: Dist | || Stor Manager | || | |+- | | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Backups of a file when it hasn't changed. | | Did the permissions change on the file or owner or group? If any of these change, TSM will take a new copy of the file as if the file had changed. -- Phillip Ford Senior Software Specialist Corporate Computer Center Schering-Plough Corp. (901) 320-4462 (901) 320-4856 FAX [EMAIL PROTECTED] -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED]] Sent: Friday, August 02, 2002 2:49 PM To: [EMAIL PROTECTED] Subject: Backups of a file when it hasn't changed. Does anyone else see this issue? I have seen several instances where TSM backed up a file when it had not been modified in anyway. Even when you look at the restore application, there are several versions of a file, each with the same create date, the same modified date, and different backed up dates. I am curious if others, or everyone sees this happening, or if I have some setting incorrect some where. There have been no always backup backups (selective) performed, so that isn't what happened. Clueless... Todd *** This message and any attachments is solely for the intended recipient. If you are not the intended recipient, disclosure, copying, use, or distribution of the information included in this message is prohibited -- please immediately and permanently delete this message.
Re: Backups of a file when it hasn't changed.
Copy mode is set to modified, and has been since day one (Nov, 2001). I assume if the copy mode was set to absolute, every file would backup every time... and that is not happening. It just appears TSM is backing things up randomly. I checked the file details using the TSM restore utility, and each version of the file was absolutely identical, with the exception of the backed up date. I see this condition Netware servers, mostly. I don't recall running across this condition on NT/W2K or my AIX servers. Todd |+-- || Rushforth, Tim| || [EMAIL PROTECTED]| || PEG.MB.CA | || Sent by: ADSM: Dist| || Stor Manager | || [EMAIL PROTECTED]| || U | || | || | || 08/02/02 03:29 PM | || Please respond to | || ADSM: Dist Stor| || Manager| || | |+-- | | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Backups of a file when it hasn't changed. | | Check that your copy group copy mode is not set to absolute. Is this windows? Are file permissions being changed? -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED]] Sent: August 2, 2002 2:49 PM To: [EMAIL PROTECTED] Subject: Backups of a file when it hasn't changed. Does anyone else see this issue? I have seen several instances where TSM backed up a file when it had not been modified in anyway. Even when you look at the restore application, there are several versions of a file, each with the same create date, the same modified date, and different backed up dates. I am curious if others, or everyone sees this happening, or if I have some setting incorrect some where. There have been no always backup backups (selective) performed, so that isn't what happened. Clueless... Todd
Re: How to save volhist and devconfig
After the backup of our database, we backup the device config and volhist files from TSM to the text files, then we perform a maksysb on our AIX server. That backup tape (4mm) goes offsite with the TSM tapes (LTO). |+- || Rob Schroeder | || robs@FAMOUSFOO| || TWEAR.COM | || Sent by: ADSM:| || Dist Stor | || Manager | || [EMAIL PROTECTED]| || ST.EDU| || | || | || 08/01/02 02:50 | || PM | || Please respond | || to ADSM: Dist | || Stor Manager | || | |+- ---| | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: How to save volhist and devconfig | ---| I have written a batch file that runs on our operators' Win2k Pro machine. The batch file maps the network drive to the TSM server and copies the required files to local diskette and another network drive. I just use the Windows scheduler to have this run every day at the same time. This may not be the best approach, but it only took 5 minutes and my operations staff doesn't have to worry about anything. Rob Schroeder Famous Footwear brian welsh brianwelsh3@HOTMTo: [EMAIL PROTECTED] AIL.COM cc: Sent by: ADSM: Subject: How to save volhist and devconfig Dist Stor Manager [EMAIL PROTECTED] .EDU 08/01/2002 02:17 PM Please respond to ADSM: Dist Stor Manager Hello, I was wondering how other administrators are saving the volhist, devconfig, dsmserv.dsk and dsmserv.opt. Every day we save the volhist and devconfig and manually ftp these files to a network-drive in case of a crash of the server. Now I'm not familiar with scripts so far, but I want to automate these job. So, I was wondering how other administrators are doing this job. Thanks for your reply. Brian. _ Chat on line met vrienden en probeer MSN Messenger uit: http://messenger.msn.nl
Re: Backups of a file when it hasn't changed.
And, how nice it was of Tivoli to set the default of that to no. That is probably my problem. I will make that change to all the Netware servers on Monday (never change things on Friday unless you have to). Thanks for that little tidbit. Todd |+-- || Rushforth, Tim| || [EMAIL PROTECTED]| || PEG.MB.CA | || Sent by: ADSM: Dist| || Stor Manager | || [EMAIL PROTECTED]| || U | || | || | || 08/02/02 04:06 PM | || Please respond to | || ADSM: Dist Stor| || Manager| || | |+-- | | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Backups of a file when it hasn't changed. | | Ahhh, Netware - check the value of the tsm option NWIGNORECOMpressbit. If set to no, then tsm will rebackup the file when netware compresses or decompresses the file. -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED]] Sent: August 2, 2002 3:20 PM To: [EMAIL PROTECTED] Subject: Re: Backups of a file when it hasn't changed. I had considered that, and honestly, I wouldn't know how to tell if the permissions had changed on the days in question. I don't see why they would have changed. There are other files in the same directory that have their only backups on the same day as the oldest backups for these files in question (the day this node started backing up to TSM). Yet some of the files have had additional backups when it doesn't appear to have been necessary. We don't normally set permissions at the file level (this is a Netware server). Plus, some of the files in question wouldn't even have a need to be accessed.. things like Ghost(R) files and old archival type documents. It really boggles the mind. Todd |+- || Ford, Phillip| || phillip.ford@S| || PCORP.COM | || Sent by: ADSM:| || Dist Stor | || Manager | || [EMAIL PROTECTED]| || ST.EDU| || | || | || 08/02/02 02:53 | || PM | || Please respond | || to ADSM: Dist | || Stor Manager | || | |+- --- -| | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Backups of a file when it hasn't changed. | --- -| Did the permissions change on the file or owner or group? If any of these change, TSM will take a new copy of the file as if the file had changed. -- Phillip Ford Senior Software Specialist Corporate Computer Center Schering-Plough Corp. (901) 320-4462 (901) 320-4856 FAX [EMAIL PROTECTED] -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED]] Sent: Friday, August 02, 2002 2:49 PM To: [EMAIL PROTECTED] Subject: Backups of a file when it hasn't changed. Does anyone else see this issue? I have seen several instances where TSM backed up a file when it had not been modified in anyway. Even when you look at the restore application, there are several versions of a file, each with the same create date, the same modified date, and different backed up dates. I am curious if others, or everyone sees this happening, or if I have some setting incorrect some where. There have been no always backup backups (selective) performed, so that isn't what happened. Clueless... Todd *** This message and any attachments is solely for the intended recipient. If you are not the intended recipient, disclosure, copying, use, or distribution of the information included in this message
Striped backup speeds.
Environment Server: TSM 4.2.1.7, AIX 4.3.3, two fiber gigabit ethernet cards, each accessing disk and tape (I know, this is bad, but it shouldn't impact this situation) Client: W2K, 4 processors (around 900mhz, I think), just over 3GB of physical memory, TSM B/A client 4.2.1.32, TSPSQL 2.2, Storage agent 4.2.1.7, three fiber gigabit ethernet cards (qLogic), one for disk, two for tape access. SAN: Shark ESS Node setup has 4 mountpoints allowed. Fiber Network: dual Brocade 2109 fiber switches. Library IBM 3584 with five LTO Ultrium fiber connected drives (three connected to one switch, two to the other switch). Utility used to measure speed during backups: qLogic Sanblade Manager. Backing up a 33GB SQL database located on the Shark to one tape, I can get about 28-30 MB/second. Backing up the same database striped to two tapes, I would get about 31 MB/second. I had the DB admin spread the three files in the database over three of the four different available drive letters (actually on 2 different packs on the ESS Shark). That made my two-stripe backup in the 36-37MB/second range (single stripe backups are still in the 28-30MB/second range). I am looking to find a way to get somewhere near the 50-60MB/s range for a backup. It doesn't make sense to stripe a backup when you are only getting a 20% increase in throughput (and this works out to less than a 20% decrease in time for the backup because of the storage agent mounting tapes, locating last files, etc, one tape at a time). I am doing testing in preparation for a 500GB database. I did some tests with very large files on each of the four drive letters, simultaneously copying the large files to nul (using four different command line prompts), to get a benchmark of data throughput on the disk fiber card. The max sustained throughput (based on the qLogic monitor) was ~66MB/second. Most of the time, the rate hovered in the 45-55 MB/second range. CPU utilization on the client node during a 2-stripe backup is around 85-95%, and during a single-stripe backup, it is around 35-45% (on all four processors). Can anyone suggest some things to check that might impact this? On another note, the backups are not completing as of my last two-stripe test. I am working with Tivoli support regarding this problem. Backups with a single stripe are working.
Re: Minimizing Database Utilization
Expiration is working fine. It starts every morning at 5 AM, and runs for about 30-40 minutes. * ANR0812I Inventory file expiration process 295 completed: examined 944182 objects, deleting 37102 backup objects, 0 archive objects, 0 DB backup volumes, and 0 recovery plan files. 0 errors were encountered * Does this mean I only have 944,182 objects being managed by the database? If so, it sounds like I do have something bloating my database. If Thomas D. can get primary and copy pool backups for 4.8 million + files in a 10GB database, and my 8GB database is filled nearly 70% utilized with less than a million objects, something is wrong somewhere. Also, I in looking for things that stand out, I found occasional instances in my 30 days worth of activity log where a TDP for SQL server started/end sessions for storage agent on the order of 5-40 times per second, sometimes lasting 4-5 seconds, sometimes lasting 4-5 minutes. I understand that anything in the activity log is in the database, too. Any ideas what could be causing the storage agent to start/end sessions 40 times per second for 5 minutes? |+ || Roger Deschner| || [EMAIL PROTECTED]| || U| || Sent by: | || ADSM: Dist | || Stor Manager | || [EMAIL PROTECTED]| || IST.EDU | ||| ||| || 07/31/02 01:14| || AM| || Please respond| || to ADSM: Dist| || Stor Manager | ||| |+ ---| | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Minimizing Database Utilization | ---| Consider the possibility that you are not expiring well enough, which could be causing database bloat. (Not to mention tape storage pool bloat.) How long does your expiration process take? Does it seem to never end? If so, you may be in an expiration bound death spiral. If you are having trouble with expiration as it is, you are in no position to take on an additional large client system, as was the start of this thread. Dramatically lengthening expiration process run times are a true sign that a TSM server is out of gas, and needs an upgrade just for its present workload, not to mention adding more work to it. Roger Deschner University of Illinois at Chicago [EMAIL PROTECTED] On Tue, 30 Jul 2002, Todd Lundstedt wrote: Well, well.. I totally read my book the wrong way. I will go recalculate. Thanks for pointing out this huge error on my part. Now I have to go figure out where the rest of my database utilization is going, too.
Re: Minimizing Database Utilization
We have no Oracle, DB2 or Exchange at the moment. All we have is SQL and Domino. TDP for SQL v 1.1 (for SQL 6.5) works that way, and we are issuing the commands to inactivate olderthan 10 days, and the management class is set to 1/0/0/0. TDP for SQL v 2.2 (for SQL 7.0/2000) and TDP for Domino correctly use the management classes/definitions assigned without the need for inactivate commands. Still looking for why some of my SQL servers using storage agent are starting and ending so many sessions per second. I am sure that is one reason why my database seems to be over utilized based on the number of files I have managed in the database. |+- || MC Matt Cooper| || (2838)| || Matt.Cooper@AMGREE| || TINGS.COM | || Sent by: ADSM:| || Dist Stor Manager | || [EMAIL PROTECTED]| || DU| || | || | || 07/31/02 09:59 AM | || Please respond to | || ADSM: Dist Stor | || Manager | || | |+- ---| | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Minimizing Database Utilization | ---| Look to make sure your TDP's are actually deleting their backups. I was burned by DB2, Oracle and MS-EXCHANGE platforms not sending the command to TSM to delete their old backups. These files are NOT really being managed by the copygroup policy the way you may think they are. The client must tell the server that the file (the TDP backups are all date/time stamped and therefore UNIQUE) has been deleted. Then your deleted file policy in the copygroup is in effect. Matt -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED]] Sent: Wednesday, July 31, 2002 10:15 AM To: [EMAIL PROTECTED] Subject: Re: Minimizing Database Utilization Expiration is working fine. It starts every morning at 5 AM, and runs for about 30-40 minutes. * ANR0812I Inventory file expiration process 295 completed: examined 944182 objects, deleting 37102 backup objects, 0 archive objects, 0 DB backup volumes, and 0 recovery plan files. 0 errors were encountered * Does this mean I only have 944,182 objects being managed by the database? If so, it sounds like I do have something bloating my database. If Thomas D. can get primary and copy pool backups for 4.8 million + files in a 10GB database, and my 8GB database is filled nearly 70% utilized with less than a million objects, something is wrong somewhere. Also, I in looking for things that stand out, I found occasional instances in my 30 days worth of activity log where a TDP for SQL server started/end sessions for storage agent on the order of 5-40 times per second, sometimes lasting 4-5 seconds, sometimes lasting 4-5 minutes. I understand that anything in the activity log is in the database, too. Any ideas what could be causing the storage agent to start/end sessions 40 times per second for 5 minutes? |+ || Roger Deschner| || [EMAIL PROTECTED]| || U| || Sent by: | || ADSM: Dist | || Stor Manager | || [EMAIL PROTECTED]| || IST.EDU | ||| ||| || 07/31/02 01:14| || AM| || Please respond| || to ADSM: Dist| || Stor Manager | ||| |+ --- | | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Minimizing Database Utilization | --- | Consider the possibility that you are not expiring well enough, which could be causing database bloat. (Not to mention
Minimizing Database Utilization
TSM 4.2.1.7 running on AIX 4.3.3 Database is 8GB, 66.2% utilized. DBVols are on an ESS shark. We have an application server that the admins would like to backup on the TSM server. The problem is, after I analyzed the backup needs of this one server, my database would grow an additional 300GB, and increase in size an additional 50-100MB each day (in actuality, it may be an additional 66% growth because I can't recall if I calculated my copy storage pool needs). They have a folder on this server (NT4.0 running a PACS system, by Fujimed, I think) that contain over 240K+ files, each file in it's own folder (so 240K+ folders) each uniquely named, never changing, and almost all of the files less than 1K in data size. There are other similar folders on this server that would significantly increase the size of the database, but not as much as this one folder. I based the increase in DB size on the 600k of database space per object stored by TSM rule. I told the admin there would be no way I could back that data up unless he created a zip file (or something similar) of all those directories/files first. He is looking into this possibility. Is there any other way to do this without increasing the size of my database more than the size of the files being backed up (what is that... 3 million times more?)? Thanks in advance for your help. Todd Lundstedt Technical Specialist Via Christi Information Management Services [EMAIL PROTECTED]
Re: Minimizing Database Utilization
Well, well.. I totally read my book the wrong way. I will go recalculate. Thanks for pointing out this huge error on my part. Now I have to go figure out where the rest of my database utilization is going, too. |+ || Thomas Denier | || Thomas.Denier@mai| || l.tju.edu| ||| || 07/30/02 10:37 AM | ||| |+ | | | | To: [EMAIL PROTECTED], [EMAIL PROTECTED] | | cc: | | Subject: Re: Minimizing Database Utilization | | I based the increase in DB size on the 600k of database space per object stored by TSM rule. I believe the rule of thumb historically given in TSM documentation is 600 bytes per object, not 600 kilobytes. I have a single client with 4.8 million backup files in one of its file systems, and several others with substantial fractions of that number. I have offsite copy pools for all backups. All of this fits in a ten gigabyte database.
Changing Storage Pools/Management Classes
I sent this out just over a month ago, and received no responses. Trying again to see if anyone can help Thanks.. Currently, we backup everything either to DISKPOOL (which migrates to LTOTAPE), or directly to LTOTAPE. Files (backup/archive client), databases (TDP for Domino, and TDP for SQL), everything is eventually stored into LTOTAPE. The databases get backed up FULL every time. The largest database we have is ~75GB. The TDP SQL databases are in a different set of management classes from the other files (TDP Domino databases have their own management class, as well). Soon, we will be creating an application that will have a 500GB database. FULL backups for SQL databases will be changing over to FULL with DIFFs (for the larger databases, FULLs on the smaller ones). I would like to segregate the SQL databases into their own tape storage pool. There will be no reclamation on the tapes in this storage pool, and collocation will be turned on, because I will be doing some striped backups. (I don't see the sense in reclaiming tapes when the major databases range in size from 10% to 400% of a tape, but give me a reason to consider it, and I will) What is the best way to handle moving/changing the storage location of my SQL databases from the LTOTAPE storage pool to the LTOSQLTAPE storage pool (I just made that name up). Should I modify the existing management classes to start using the new storage pool? If so, what happens to the files in the current storage pool? Should I create a new management class pointing to the new storage pool and change the INCLUDE statements to point to the new management class? If so, what happens to the files in the current storage pool and the current management class? Is there another way to do this? Thanks in advance for your help. Todd Lundstedt Technical Specialist Via Christi Information Management Services [EMAIL PROTECTED] ps FYI. Since the first time I sent this note out, I have done some testing. After working with TSM support on an issue that came up in my testing, I find out that a management class cannot be changed if you are backing up LANFREE. You must back up the file across the LAN in order to change the management class. Once the active version of the file is in the new management class, LANFREE backups can resume. However, I am still getting occasional errors when I do this, and then try to perform striped backups.
Re: Changing Storage Pools/Management Classes
Thanks, Wanda. I think you answered my question. I just want to make sure there are no issues. So, to clarify... I have had management classes setup to handle these files for a long time, and they are pointing to either a diskpool or a tapepool. I want all of the files in these management classes to eventually be stored in a different disk and tape pool. It sounds like the best way to do this is (very simply) to just change the copy-to destination of the existing management classes for these files to point to the new storage pools. I will still maintain the original storage pools for other management classes still in use. I just wasn't sure of any negative aspects of this change. Do you see any gotchas? anything in particular of which I should be aware? Thanks again, Todd |+-- || Prather, Wanda| || Wanda.Prather@J| || HUAPL.EDU | || Sent by: ADSM: | || Dist Stor | || Manager| || [EMAIL PROTECTED]| || T.EDU | || | || | || 07/02/02 11:50 | || AM | || Please respond | || to ADSM: Dist | || Stor Manager | || | |+-- | | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Changing Storage Pools/Management Classes | | If you want some files in a different tape storage pool, you will need a new management class for them. If you are backing up direct to tape, just point the new management class to the new tape storage pool. Use INCLUDE statements in dsm.opt to bind just the files you want to the new management class. If you are backing up to disk first, you will need to create a new disk pool for these files as well as a separate tape pool. (You can't have one primary disk pool that migrates to two different tape pools.) Make the new management class point to the new disk pool and have the disk pool point to the new tape pool. When you backup files to the new management class, the NEW backup versions will go to the new pool(s). Nothing happens to older/existing backup versions of those files. They stay on whatever media they were on originally.Nothing triggers them to move or relocate. They aren't hurt, you can still restore from them if you need to. The best thing to do is just leave them where they sit, and let them expire when they normally would. -Original Message- From: Todd Lundstedt [mailto:[EMAIL PROTECTED]] Sent: Tuesday, July 02, 2002 11:48 AM To: [EMAIL PROTECTED] Subject: Changing Storage Pools/Management Classes I sent this out just over a month ago, and received no responses. Trying again to see if anyone can help Thanks.. Currently, we backup everything either to DISKPOOL (which migrates to LTOTAPE), or directly to LTOTAPE. Files (backup/archive client), databases (TDP for Domino, and TDP for SQL), everything is eventually stored into LTOTAPE. The databases get backed up FULL every time. The largest database we have is ~75GB. The TDP SQL databases are in a different set of management classes from the other files (TDP Domino databases have their own management class, as well). Soon, we will be creating an application that will have a 500GB database. FULL backups for SQL databases will be changing over to FULL with DIFFs (for the larger databases, FULLs on the smaller ones). I would like to segregate the SQL databases into their own tape storage pool. There will be no reclamation on the tapes in this storage pool, and collocation will be turned on, because I will be doing some striped backups. (I don't see the sense in reclaiming tapes when the major databases range in size from 10% to 400% of a tape, but give me a reason to consider it, and I will) What is the best way to handle moving/changing the storage location of my SQL databases from the LTOTAPE storage pool to the LTOSQLTAPE storage pool (I just made that name up). Should I modify the existing management classes to start using the new storage pool? If so, what happens to the files in the current storage pool? Should I create a new management class
What would be the SQL select to determine....
... a list of files (with directory names) in a storage pool that only have one version? Or can you do that with an SQL select? I am trying to identify files with unique names that may be of a temporary nature, so I can point them to a different management class with a Retain Only Version count of zero. Thanks in advance, Todd
Changing Storage Pools/Management Classes
Currently, we backup everything either to DISKPOOL (which migrates to LTOTAPE), or directly to LTOTAPE. Files (backup/archive client), databases (TDP for Domino, and TDP for SQL), everything is eventually stored into LTOTAPE. The databases get backed up FULL every time. The largest database we have is ~75GB. The TDP SQL databases are in a different set of management classes from the other files (TDP Domino databases have their own management class, as well). Soon, we will be creating an application that will have a 500GB database. FULL backups for SQL databases will be changing over to FULL with DIFFs (for the larger databases, FULLs on the smaller ones). I would like to segregate the SQL databases into their own tape storage pool. There will be no reclamation on the tapes in this storage pool, and collocation will be turned on, because I will be doing some striped backups. (I don't see the sense in reclaiming tapes when the major databases range in size from 10% to 400% of a tape, but give me a reason to consider it, and I will) What is the best way to handle moving/changing the storage location of my SQL databases from the LTOTAPE storage pool to the LTOSQLTAPE storage pool (I just made that name up). Should I modify the existing management classes to start using the new storage pool? If so, what happens to the files in the current storage pool? Should I create a new management class pointing to the new storage pool and change the INCLUDE statements to point to the new management class? If so, what happens to the files in the current storage pool and the current management class? Is there another way to do this? Thanks in advance for your help. Todd Lundstedt Technical Specialist Via Christi Information Management Services [EMAIL PROTECTED]
TDP for SQL and Stripes
I am starting to play with the stripe function of the TDP for SQL. I got to wondering... If you have a database backed up directly to tape using 4 stripes... you need 4 tape drives to do that.. and, 4 tape drives to do a restore. I got that much... But.. What happens if, by reclamation process, two or more of the striped portions of the backup get reclaimed onto the same tape? Will TSM allow that to happen? if so, how do you do a restore? Any other gotchya's out there?
Re: Expiring specific files
Wow! Steve, that sounds like.. um.. fun?! I follow along with what you are trying to do, but I am not sure what this would do to the application, since that application watches this directory (as, I am guessing, a sort of queue). I would hesitate to create any file on my own in this directory, even more so, a file that has already been created by the application in this directory. I will certainly keep this workaround in mind for future issues, though. Just to clarify.. is there no way to expire a single file version, active or inactive? |+ || Steve Harris | || [EMAIL PROTECTED]| || LD.GOV.AU| || Sent by: ADSM: Dist | || Stor Manager | || [EMAIL PROTECTED]| ||| ||| || 04/23/02 06:59 PM | || Please respond to | || ADSM: Dist Stor | || Manager | ||| |+ | | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Expiring specific files | | Todd, I just had this problem with a bunch of oracle archive files that had been placed into the wrong directory and so had picked up the default retention (retonly 365) instead of the oracle one (retonly 35). Your suggested approach will work fine for active files, which will be rebound, but inactive files will not be. As you have suggested, set up your includes to give these files a short retention period. Then, generate your list of inactive files. In my case it was easy as the files had ascending sequence numbers. If your case isn't that simple you could use either the ba client q backup command or alternatively use the admin client to select appropriately from the contents table and post-process the output. Both approaches have issues, but are do-able. Take this list and process it through your favourite scripting language. In batches of say 100 files, create a new zero length file with the name of the file you want deleted from tsm. Under unix I'd use touch, under windows I'm not sure. back up the directory delete the batch Repeat until all files processed. back up the directory one last time Set your include/excludes to whatever your long term requirement is for this. This should fix the problem. Regards Steve Harris AIX and TSM Admin Queensland Health, Brisbane Australia [EMAIL PROTECTED] 24/04/2002 6:32:23 I have a bunch of files that have been backed up in one directory on this node that we no longer need to keep. These files all have unique names, but share the extension of .prt. There is a very large number of inactive versions (simply because these are temporary files that get deleted each day). I would like to find a way to expire the files in this folder. Is there a command to do this? or do I have to change the management class for this file to a dummy management class/backup copy group (say, nohold set to 1, 0, 0, 0)? If I have to change the management class, what are the steps I need to do to accomplish this? Is it simply... Create new management class/backup copy group (call it nohold)? modify the options file for this node to have an include path\*.prt nohold and then run an incremental Is that basically it? ** This e-mail, including any attachments sent with it, is confidential and for the sole use of the intended recipient(s). This confidentiality is not waived or lost if you receive it and you are not the intended recipient(s), or if it is transmitted/ received in error. Any unauthorised use, alteration, disclosure, distribution or review of this e-mail is prohibited. It may be subject to a statutory duty of confidentiality if it relates to health service matters. If you are not the intended recipient(s), or if you have received this e-mail in error, you are asked to immediately notify the sender by telephone or by return e-mail. You should also delete this e-mail message and destroy any hard copies produced. **
Re: SQL Output
If you are doing this from the command line in AIX. dsmadmc -id=yourID -password=yourPW select whatever from whereever, etc | tail +XX | grep -v YY new_filename If you are doing this to a file that was created from the dsmadmc client already.. it is very similar... cat filename | tail +XX | grep -v YY new_filename where XX = the number of rows you want removed from the top where YY = some unique text in the line you want to exclude David E Ehresman To: [EMAIL PROTECTED] deehre01@LOUIS cc: VILLE.EDUSubject: SQL Output Sent by: ADSM: Dist Stor Manager [EMAIL PROTECTED] ST.EDU 04/24/02 12:15 PM Please respond to ADSM: Dist Stor Manager I there a way to get JUST the output from a select statement into a file without the TSM version header info at the beginning and the condition code at the end? This is TSM 4.2 on Aix 5.1. David
Expiring specific files
I have a bunch of files that have been backed up in one directory on this node that we no longer need to keep. These files all have unique names, but share the extension of .prt. There is a very large number of inactive versions (simply because these are temporary files that get deleted each day). I would like to find a way to expire the files in this folder. Is there a command to do this? or do I have to change the management class for this file to a dummy management class/backup copy group (say, nohold set to 1, 0, 0, 0)? If I have to change the management class, what are the steps I need to do to accomplish this? Is it simply... Create new management class/backup copy group (call it nohold)? modify the options file for this node to have an include path\*.prt nohold and then run an incremental Is that basically it?
Re: Expiring specific files
Nope.. sorry .. I didn't include any specific info.. TSM Server 4.2.1.7 running on AIX 433 the node in question is Netware 4.11 running TSM B/A Client for Netware 4.1.3.0. It is an application and fileserver, so no, there is a LOT more stuff on this filespace than just the files I want to expire off. Todd |+-- || David Longo | || David.Longo@HEALTH-| || FIRST.ORG | || Sent by: ADSM: Dist| || Stor Manager | || [EMAIL PROTECTED]| || U | || | || | || 04/23/02 03:55 PM | || Please respond to | || ADSM: Dist Stor| || Manager| || | |+-- | | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: Expiring specific files | | If this was a unix platform AND the files are the ONLY thing in a filesystem (filespace) then just use DELETE FILESPACE on the TSM server would be simplest way. David Longo [EMAIL PROTECTED] 04/23/02 04:32PM I have a bunch of files that have been backed up in one directory on this node that we no longer need to keep. These files all have unique names, but share the extension of .prt. There is a very large number of inactive versions (simply because these are temporary files that get deleted each day). I would like to find a way to expire the files in this folder. Is there a command to do this? or do I have to change the management class for this file to a dummy management class/backup copy group (say, nohold set to 1, 0, 0, 0)? If I have to change the management class, what are the steps I need to do to accomplish this? Is it simply... Create new management class/backup copy group (call it nohold)? modify the options file for this node to have an include path\*.prt nohold and then run an incremental Is that basically it? MMS health-first.org made the following annotations on 04/23/02 17:09:31 -- This message is for the named person's use only. It may contain confidential, proprietary, or legally privileged information. No confidentiality or privilege is waived or lost by any mistransmission. If you receive this message in error, please immediately delete it and all copies of it from your system, destroy any hard copies of it, and notify the sender. You must not, directly or indirectly, use, disclose, distribute, print, or copy any part of this message if you are not the intended recipient. Health First reserves the right to monitor all e-mail communications through its networks. Any views or opinions expressed in this message are solely those of the individual sender, except (1) where the message states such views or opinions are on behalf of a particular entity; and (2) the sender is authorized by the entity to give such views or opinions. ==
Re: How Best to Suspend a Schedule
Wouldn't this disable ALL schedules, instead of just the one you need to suspend? |+ || Edgardo Moso | || edgardo_moso@KINDREDHEALT| || HCARE.COM| || Sent by: ADSM: Dist Stor | || Manager | || [EMAIL PROTECTED]| ||| ||| || 04/19/02 12:22 PM | || Please respond to ADSM: | || Dist Stor Manager| ||| |+ | | | | To: [EMAIL PROTECTED] | | cc: | | Subject: Re: How Best to Suspend a Schedule | | For #1 you can set , disablesched Yes in the dsmserv.opt file. From: Kilmer, Paul [EMAIL PROTECTED] on 04/19/2002 09:02 AM Please respond to ADSM: Dist Stor Manager [EMAIL PROTECTED] To: [EMAIL PROTECTED] cc: Subject: How Best to Suspend a Schedule As a relative newbie to TSM, I'd be interested to hear what is the easiest way to temporarily: 1- Suspend a backup schedule from running. 2- Suspend an individual client's participation in a backup schedule. I've accomplished #2 by removing the client schedule association, though I don't regard this method to be particularly elegant. Thanks in advance. Paul E. Kilmer Lead Technical Architect New World Pasta [EMAIL PROTECTED] (717) 526-2431 || |*http://www.eteamz.com/cdboys www.eteamz.com/cdboys *| || ||