Allen, 

Your post got me thinking as to just why I decided what I did  ―  and now I remember :)

There was another management requirement that all production servers be backed up in 
full once per year and that snapshot be kept "forever" - there is no reasoning with 
this, its one of those stupid mandates that applies to the whole of the state 
government, and if data is not able to be categorized, then it must be kept.

I think you'll recognise that having a third TSM Server for the yearly backup isn't 
really an option, so an archive mechanism is the only one that will work for that.  
Backupsets weren't an option when this was set up as they were very new and not mature 
(IMHO they're still not mature because I can't stack them, track them in a reasonable 
way, or use them for TDP data).  Having an archive mechanism for monthlies means just 
changing the archive mgmtclass once a year and voila, your monthly becomes a yearly.

Given that management loves the idea of permanent retention, database size is 
obviously not a problem.  Hence the argument comes down to ease of maintenance and the 
monthly archive is better at that - I have a perl script that generates the backup 
schedules with correct domain statements based on the current backup filespaces every 
month, and each February we generate with the "eternal" mgmtclass.  Its mostly 
automated.

Steve.


>>> [EMAIL PROTECTED] 17/07/2004 2:07:39 >>>
==> In article <[EMAIL PROTECTED]>, Steve Harris <[EMAIL PROTECTED]> writes:


> at 1% , 1-(0.99**30), or about .25
> at 2%,  1-(0.98**30) , or about .45
> at 3%,  1-(0.97**30), or about .60   (Please feel free to correct my maths if I'm 
> wrong - probability was never my strong point)

I think your math is good; I'd add though that there's a -GREAT- deal of
locality of reference: Or in other words, if 3% of your files change a day,
then for user filespace probably 95% of the files that change tomorrow will be
the files that chaged yesterday, and so on.  I think that 25 - 30% for userdir
space is probably about right.


> Now, of course its often the same files which change day after day, so real
> experience should be better than this, but at the time, I decided that the
> overhead of mainitianing two TSMs (and two clients per node) wasn't worth the
> benefit, and went with archives.

But I would disagree with your logic;

In place of the incremental possibilities, which would have led you at worst
above to backing up 60% every month, you're choosing to back up 100% every
month, without fail.

I think that this represents a substantially more costly strategy.

In fact, given the monthly-for-five-years figure and your numbers above, it's
about twice as expensive in facilities to archive monthly than to run
incrementals with similar retention characteristics; If my guess is closer to
correct, it's three- to four- times the cost.

For a small amount of data, this may be cheaper than the human organizational
time to build two sets of nodes; but by the time you get even medium size, I
think it would be pretty expensive.

- Allen S. Rout



***********************************************************************************
This email, including any attachments sent with it, is confidential and for the sole 
use of the intended recipient(s).  This confidentiality is not waived or lost, if you 
receive it and you are not the intended recipient(s), or if it is transmitted/received 
in error.

Any unauthorised use, alteration, disclosure, distribution or review of this email is 
prohibited.  It may be subject to a statutory duty of confidentiality if it relates to 
health service matters.

If you are not the intended recipient(s), or if you have received this email in error, 
you are asked to immediately notify the sender by telephone or by return email.  You 
should also delete this email and destroy any hard copies produced.
***********************************************************************************

Reply via email to