Thanks a lot, Brock, for your comprehensive post and also to the others. I 
haven't fully worked through your example cases yet, but it  will certainly 
help me to get my head around it all. Maybe it helps if I provide a few 
more details about how the data/images are organized:

I run a Linux based virtualization cluster on RAID6-hosts with Windows VMs. 
The images are organised in windows-folders of 2TB size each, like 
"E:\img\img01\" to currently "E:\img\img17\". 
Once a folder is full, it's contents will never change again. They're like 
archives that will be read from but no more written to. 

So I thought I'd proceed like this:
1. Backup "img01" to "img17" to tape, store the tapes offsite. 
2. Do this a second time and store the tapes offsite, seperate from the 
first generation.
3. Do this a third time to disk, for quick access if needed.
4. Make sure the catalog of at least 1. and 2. is in a very safe place.
5. Develop a daily backup strategy - starting with "img18".

As for  (1.) - (3.) I have created seperate Full-jobs for each 
imgXX-folder. (1.) has already completed successfully, (2.) is currently in 
progress.
I thought that once (1.) and (2.) are completed successfully I'm safe what 
"img01-17" is concerned and never have to consider these folders for backup 
again. Right or am I missing something?

What I'd like to discuss here is (5.) - under consideration of a few 
parameters:
- the daily increment of image data is roughly 50 GB. BTW: The images 
(DICOM, JPEG2000) don't compress at all :). 
- for legal reasons we have to store the images on WORM-media. So I need a 
daily job that writes to tape.
- the doctors want best possible protection against fire, supernova, 
Borg-attack etc. They want a daily tape change routine with the latest 
WORM-tape taken offsite. 

For the daily tape change I could buy another LTO drive. I can also expand 
my backup-server to fit above (3.) and the daily increment.

So, here's what I thought I need to develop:
- Backup the daily increment to disk.
- Backup the same daily increment to a WORM tape (in a new 1-slot drive) 
that is part of a "daily change" pool of tapes (MON-SAT or so...)
- Append the same daily increment to another WORM tape in the 8-slot 
loader. Once the tape is full, take it offsite and add a fresh tape in the 
loader.

If that strategy doesn't sound to weird I need to transfer this into a 
working bareos config.
Sorry if it all sounds confusing but for me it's still really, really 
complex.

Thanks
Andreas

bro...@mlds-networks.com schrieb am Mittwoch, 5. August 2020 um 20:21:10 
UTC+2:

> You will have some complexity with the size of your data and the size of 
> your loader. Unless your data compresses really well.
> Does it have more than one tape drive? Your total loader capacity is 48 
> TBytes raw, and you need 2x your full size to do Consolidations or new 
> Fulls or you have gaps in your protection.
>
> If I’m reading this right you want an off site copy.
>
> If that’s correct I would go about this two different ways, 
>
> * Get a much bigger loader with 2 drives
> or
> * Expand backups server raid6 to have Full + growth*Growth Window capacity 
>
> I would then use migrate+Archive jobs to make my off site and copy to tape.
>
> In the first case you can avoid the extra migrate, just do an archive to a 
> pool of tapes you eject.
>
> Workflow case 1 Bigger Tape loader 2 or more tape drives.
> * Spool to Disk
> * AI-Consolidated Pool Tape
> * AI-Incremental Pool Disk
> * Offsite Pool Tape
>
> Fulls and Consolidations backups go to AI-Consolidated Tape pool,
> Your daily go to disk until they are consolidated into tape.
>
> To create your off sites you can use a copy job of whatever full you want.
>
> https://docs.bareos.org/TasksAndConcepts/AlwaysIncrementalBackupScheme.html#copy-jobs
>
> I personally for offsite to avoid issues with Always Incremental jobs use 
> an Archive job
>
> https://docs.bareos.org/TasksAndConcepts/AlwaysIncrementalBackupScheme.html#virtual-full-jobs
>
> This avoids the offsite tapes being upgraded to the primary copy when the 
> Consolidate job prunes the older jobs.
>
> To do a VirtualFull archive job like this though you need enough tape for 
> 2x the data otherwise your swapping tapes every 5.5 hours.
> I would then use a script called at the end of the archive job to eject 
> the volumes and make volumestatus=Used from that Offsite pool. Load new 
> tapes label and put back in Offsite for the next archive.
>
> To do AI jobs this way and the VirtualFull archive job you need 2 tape 
> drives to read from one pool/tape and write to the other. It will be fast 
> as you get LTO speeds. No second Full backup.
>
>
>
> Workflow case 2 Bigger Disk in Backup Server
> * Spool if you want to avoid failed jobs cluttering disk volumes, maybe 
> skip.
> * AI-Consolidated Pool Disk
> * AI-Incremental Pool Disk
> * Offsite Pool Tape
>
> This would work similar, but your Fulls and Consolidations all happen on 
> Disk. This now means your disk needs to be 2x Full + Incremental. As your 
> Full or VirtualFull consolidation will read all the data and make a second 
> copy before purging the old copy from the catalog. So you won’t need the 
> disk space long, but you will need it.
>
> As for Offsite I would do the archive job again, Make a VirtualFull mark 
> it as archive to the Offsite pool in the Tape loader. Eject all 7-8 tapes 
> and load new ones.
>
> Assuming you don’t want to be doing a new full Offsite every time, you can 
> do copy jobs from your pools but I have found Bareos to behave in an 
> unintuitive way when it comes Consolidations and offline media Copies. You 
> end up having to run the consolidate job twice (once primary copy, Copy job 
> then becomes primary and is picked up in next consolidation, where it wants 
> to read from that media now) When you really just want the Copy jobs to be 
> purged in most cases along with the primary copy on a Consolidate. 
>
> So I hoped you wanted just a full offsite say 1/month.
>
> Another option if you wanted is to setup a second SD with disk off site, 
> and send just your incremental there as part of a Copy job. Sneakernet the 
> tapes with Full less often.
>
>
> In general Always Incremental with copies for off site needs some work 
> unless you are ok having a full snapshot every so often with an Archive 
> job. I really hope Bareos improves on this in the future.
>
>
> Lastly if you do use the archive job. It’s it’s own stand alone job, As 
> your using WORM and cannot reuse your tapes, you probably want to crank up 
> their retention so you can easly restore files form any batch of tapes. I 
> would also for all of these run an extra backup of your catalog and append 
> it to the offsite copies so you can pull catalog data in the event 
> everything is lost. If you are encrypting (you are right?) be sure to keep 
> a second copy of all your keys, I keep mine in OnePassword. I also upload a 
> dump of the catalog to a cloud storage provider (Secure as you will). Doing 
> a bscan of 35 TBytes will take a while, so please please please keep extra 
> copies of your catlog and keys. 
>
> Also do a lot of testing with the archive jobs. How to turn them back into 
> backup jobs to make a new VirtualFull for the real job, etc. If your Full’s 
> ever screw up, you will want to do this (what I do for road warriors where 
> Fulls are not really possible) to avoid the long time to do a new full from 
> the source. Putting an Archive job back into a backup job, running a manual 
> virtual full of a job will suck the files defined in the archive job in 
> (won’t if it’s an archive, thus the change to update job type=B) and 
> rebuild a full, then do an incremental.
>
>
> I personally do a version of case 2, but I use Migrate jobs to move the 
> AI-Consolidated to a tape pool, but my jobs full’s are much smaller than 
> yours. I also only do an offsite archive job once a month. Otherwise 
> Incremental and AI-Consolidated on Disk, Migrate to tape pool purge 
> AI-Consolidated jobs quickly. So when I consolidate it’s Tape + Disk —> 
> Disk —> Migrate Tape —> Archive Tape
>
> # copy job to long term tape
> Job {
> Name = "Migrate-To-Offsite-AI-Consolidated"
> Client = myth-fd
> Type = Migrate
> Purge Migration Job = yes
> Pool = AI-Consolidated
> Level = Full
> Next Pool = Offsite
> Schedule = WeeklyCycleAfterBackup
> Allow Duplicate Jobs = no
> Priority = 4 #before catalog dump
> Messages = Standard
> Selection Type = PoolTime # 7 days
> Spool Data = No
> Selection Pattern = "."
> RunAfterJob = "/usr/local/bin/prune.sh”
> }
>
> The script forces the volumes to truncate/prune after I move the jobs off.
>
>
> Brock Palen
> 1 (989) 277-6075 <(989)%20277-6075>
> bro...@mlds-networks.com
> www.mlds-networks.com
> Websites, Linux, Hosting, Joomla, Consulting
>
>
>
> > On Aug 5, 2020, at 5:40 AM, a.br...@mail.de <a.br...@mail.de> wrote:
> > 
> > 
> > Hello everybody,
> > 
> > I'm in the process of developing a regular backup-strategy and found 
> that I need some assistance. Here are the parameters in short:
> > - 35TB of medical imaging data
> > - daily increment of 50-60GB
> > - one site, 10Gb/s Backbone
> > - Overland NEOs LTO7 Storageloader, 8-bay.
> > attached to
> > - dedicated backupserver with 20TB RAID6, will be enhanced as needed
> > 
> > I have already backed up all data on LTO7 tape (WORM, for legal reasons) 
> as of Dec'19.
> > A 2nd generation, also LTO7 WORM, is currently in progress (VERY! slow, 
> ~12MB/s, different story). Tapes are/will be stored offsite.
> > After that I'm planning to do a 3rd generation on disk inhouse and amend 
> the 1st and 2nd gen on tape so that I end up with three generations 
> identical to a certain cutoff date.
> > 
> > Then, what to do next? How could a daily routine look like? 
> > The radiologists are very concerned about their data and would like to 
> see a daily tape change with the ejected tape being taken offsite in the 
> evening. A 1-bay LTO8 drive could be purchased then, the daily tape change 
> would be done by an apprentice or so...
> > 
> > So I thought about an Always-Incremental-B2D2T-strategy starting with 
> the above cutoff day. But I still have too less experience and so I'm 
> struggeling hard to develop a clear structure in my head - WHAT is WHEN 
> being copied WHERE - let alone transform that into a working bareos 
> configuration.
> > 
> > Do my thoughts appear reasonable up to that point?
> > BTW: can a daily tape change be realized at all, where you just push the 
> button to eject the TUE-tape, insert the WED-tape, and so on without having 
> to stop the storage-daemon in order to unlock the tape?
> > 
> > Thanks for helping me with the first steps to bring this under way.
> > 
> > Andreas
> > 
> > 
> > 
> > -- 
> > You received this message because you are subscribed to the Google 
> Groups "bareos-users" group.
> > To unsubscribe from this group and stop receiving emails from it, send 
> an email to bareos-users...@googlegroups.com.
> > To view this discussion on the web visit 
> https://groups.google.com/d/msgid/bareos-users/64fe4cfa-6c87-4d28-aed9-9e30de285ee1n%40googlegroups.com
> .
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to bareos-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/bareos-users/4aa25985-fd74-4e2b-bb71-95ed37154bc1n%40googlegroups.com.

Reply via email to