Thanks a lot, Brock, for your comprehensive post and also to the others. I 
haven't fully worked through your example cases yet, but it  will certainly 
help me to get my head around it all. Maybe it helps if I provide a few 
more details about how the data/images are organized:

I run a Linux based virtualization cluster on RAID6-hosts with Windows VMs. 
The images are organised in windows-folders of 2TB size each, like 
"E:\img\img01\" to currently "E:\img\img17\". 
Once a folder is full, it's contents will never change again. They're like 
archives that will be read from but no more written to. 

So I thought I'd proceed like this:
1. Backup "img01" to "img17" to tape, store the tapes offsite. 
2. Do this a second time and store the tapes offsite, seperate from the 
first generation.
3. Do this a third time to disk, for quick access if needed.
4. Make sure the catalog of at least 1. and 2. is in a very safe place.
5. Develop a daily backup strategy - starting with "img18".

As for  (1.) - (3.) I have created seperate Full-jobs for each 
imgXX-folder. (1.) has already completed successfully, (2.) is currently in 
progress.
I thought that once (1.) and (2.) are completed successfully I'm safe what 
"img01-17" is concerned and never have to consider these folders for backup 
again. Right or am I missing something?

What I'd like to discuss here is (5.) - under consideration of a few 
parameters:
- the daily increment of image data is roughly 50 GB. BTW: The images 
(DICOM, JPEG2000) don't compress at all :). 
- for legal reasons we have to store the images on WORM-media. So I need a 
daily job that writes to tape.
- the doctors want best possible protection against fire, supernova, 
Borg-attack etc. They want a daily tape change routine with the latest 
WORM-tape taken offsite. 

For the daily tape change I could buy another LTO drive. I can also expand 
my backup-server to fit above (3.) and the daily increment.

So, here's what I thought I need to develop:
- Backup the daily increment to disk.
- Backup the same daily increment to a tape (in a new 1-slot drive) that is 
part of a "daily change" pool of tapes (MON-SAT or so...)
- Append the same daily increment to another WORM tape in the 8-slot 
loader. Once the tape is full, take it offsite and add a fresh tape in the 
loader.

If that strategy doesn't sound to weird I need to transfer this into a 
working bareos config.
Sorry if it all sounds confusing but for me it's still really, really 
complex.

Thanks
Andreas



spadaj...@gmail.com schrieb am Freitag, 7. August 2020 um 09:56:48 UTC+2:

> As I wrote earlier, this looks more like archiving plan, not a backup one 
> (or a combination of backup and archiving). But more to the point - in case 
> of backups you have to have a verification plan and periodical restore 
> tests. In case of archiving you need to have a verification plan (i.e. 
> every half a year you read each archive unit, check whether it's readable 
> and if its checksum is correct; in case of more critical data you might 
> want to even have some kind of functional tests like trying to read the 
> data into appropriate software and check whether the data is still 
> parseable and loadable; If any copy fails you'd want to re-create it from 
> other existing copies).
> On 07.08.2020 09:49, a.br...@mail.de wrote:
>
> Thanks a lot, Brock, for your comprehensive post and also to the others. I 
> haven't fully worked through your example cases yet, but it  will certainly 
> help me to get my head around it all. Maybe it helps if I provide a few 
> more details about how the data/images are organized:
>
> I run a Linux based virtualization cluster on RAID6-hosts with Windows 
> VMs. 
> The images are organised in windows-folders of 2TB size each, like 
> "E:\img\img01\" to currently "E:\img\img17\". 
> Once a folder is full, it's contents will never change again. They're like 
> archives that will be read from but no more written to. 
>
> So I thought I'd proceed like this:
> 1. Backup "img01" to "img17" to tape, store the tapes offsite. 
> 2. Do this a second time and store the tapes offsite, seperate from the 
> first generation.
> 3. Do this a third time to disk, for quick access if needed.
> 4. Make sure the catalog of at least 1. and 2. is in a very safe place.
> 5. Develop a daily backup strategy - starting with "img18".
>
> As for  (1.) - (3.) I have created seperate Full-jobs for each 
> imgXX-folder. (1.) has already completed successfully, (2.) is currently in 
> progress.
> I thought that once (1.) and (2.) are completed successfully I'm safe what 
> "img01-17" is concerned and never have to consider these folders for backup 
> again. Right or am I missing something?
>
> What I'd like to discuss here is (5.) - under consideration of a few 
> parameters:
> - the daily increment of image data is roughly 50 GB. BTW: The images 
> (DICOM, JPEG2000) don't compress at all :). 
> - for legal reasons we have to store the images on WORM-media. So I need a 
> daily job that writes to tape.
> - the doctors want best possible protection against fire, supernova, 
> Borg-attack etc. They want a daily tape change routine with the latest 
> WORM-tape taken offsite. 
>
> For the daily tape change I could buy another LTO drive. I can also expand 
> my backup-server to fit above (3.) and the daily increment.
>
> So, here's what I thought I need to develop:
> - Backup the daily increment to disk.
> - Backup the same daily increment to a WORM tape (in a new 1-slot drive) 
> that is part of a "daily change" pool of tapes (MON-SAT or so...)
> - Append the same daily increment to another WORM tape in the 8-slot 
> loader. Once the tape is full, take it offsite and add a fresh tape in the 
> loader.
>
> If that strategy doesn't sound to weird I need to transfer this into a 
> working bareos config.
> Sorry if it all sounds confusing but for me it's still really, really 
> complex.
>
> Thanks
> Andreas
>
>
> bro...@mlds-networks.com schrieb am Mittwoch, 5. August 2020 um 20:21:10 
> UTC+2:
>
>> You will have some complexity with the size of your data and the size of 
>> your loader. Unless your data compresses really well. 
>> Does it have more than one tape drive? Your total loader capacity is 48 
>> TBytes raw, and you need 2x your full size to do Consolidations or new 
>> Fulls or you have gaps in your protection. 
>>
>> If I’m reading this right you want an off site copy. 
>>
>> If that’s correct I would go about this two different ways, 
>>
>> * Get a much bigger loader with 2 drives 
>> or 
>> * Expand backups server raid6 to have Full + growth*Growth Window 
>> capacity 
>>
>> I would then use migrate+Archive jobs to make my off site and copy to 
>> tape. 
>>
>> In the first case you can avoid the extra migrate, just do an archive to 
>> a pool of tapes you eject. 
>>
>> Workflow case 1 Bigger Tape loader 2 or more tape drives. 
>> * Spool to Disk 
>> * AI-Consolidated Pool Tape 
>> * AI-Incremental Pool Disk 
>> * Offsite Pool Tape 
>>
>> Fulls and Consolidations backups go to AI-Consolidated Tape pool, 
>> Your daily go to disk until they are consolidated into tape. 
>>
>> To create your off sites you can use a copy job of whatever full you 
>> want. 
>>
>> https://docs.bareos.org/TasksAndConcepts/AlwaysIncrementalBackupScheme.html#copy-jobs
>>  
>>
>> I personally for offsite to avoid issues with Always Incremental jobs use 
>> an Archive job 
>>
>> https://docs.bareos.org/TasksAndConcepts/AlwaysIncrementalBackupScheme.html#virtual-full-jobs
>>  
>>
>> This avoids the offsite tapes being upgraded to the primary copy when the 
>> Consolidate job prunes the older jobs. 
>>
>> To do a VirtualFull archive job like this though you need enough tape for 
>> 2x the data otherwise your swapping tapes every 5.5 hours. 
>> I would then use a script called at the end of the archive job to eject 
>> the volumes and make volumestatus=Used from that Offsite pool. Load new 
>> tapes label and put back in Offsite for the next archive. 
>>
>> To do AI jobs this way and the VirtualFull archive job you need 2 tape 
>> drives to read from one pool/tape and write to the other. It will be fast 
>> as you get LTO speeds. No second Full backup. 
>>
>>
>>
>> Workflow case 2 Bigger Disk in Backup Server 
>> * Spool if you want to avoid failed jobs cluttering disk volumes, maybe 
>> skip. 
>> * AI-Consolidated Pool Disk 
>> * AI-Incremental Pool Disk 
>> * Offsite Pool Tape 
>>
>> This would work similar, but your Fulls and Consolidations all happen on 
>> Disk. This now means your disk needs to be 2x Full + Incremental. As your 
>> Full or VirtualFull consolidation will read all the data and make a second 
>> copy before purging the old copy from the catalog. So you won’t need the 
>> disk space long, but you will need it. 
>>
>> As for Offsite I would do the archive job again, Make a VirtualFull mark 
>> it as archive to the Offsite pool in the Tape loader. Eject all 7-8 tapes 
>> and load new ones. 
>>
>> Assuming you don’t want to be doing a new full Offsite every time, you 
>> can do copy jobs from your pools but I have found Bareos to behave in an 
>> unintuitive way when it comes Consolidations and offline media Copies. You 
>> end up having to run the consolidate job twice (once primary copy, Copy job 
>> then becomes primary and is picked up in next consolidation, where it wants 
>> to read from that media now) When you really just want the Copy jobs to be 
>> purged in most cases along with the primary copy on a Consolidate. 
>>
>> So I hoped you wanted just a full offsite say 1/month. 
>>
>> Another option if you wanted is to setup a second SD with disk off site, 
>> and send just your incremental there as part of a Copy job. Sneakernet the 
>> tapes with Full less often. 
>>
>>
>> In general Always Incremental with copies for off site needs some work 
>> unless you are ok having a full snapshot every so often with an Archive 
>> job. I really hope Bareos improves on this in the future. 
>>
>>
>> Lastly if you do use the archive job. It’s it’s own stand alone job, As 
>> your using WORM and cannot reuse your tapes, you probably want to crank up 
>> their retention so you can easly restore files form any batch of tapes. I 
>> would also for all of these run an extra backup of your catalog and append 
>> it to the offsite copies so you can pull catalog data in the event 
>> everything is lost. If you are encrypting (you are right?) be sure to keep 
>> a second copy of all your keys, I keep mine in OnePassword. I also upload a 
>> dump of the catalog to a cloud storage provider (Secure as you will). Doing 
>> a bscan of 35 TBytes will take a while, so please please please keep extra 
>> copies of your catlog and keys. 
>>
>> Also do a lot of testing with the archive jobs. How to turn them back 
>> into backup jobs to make a new VirtualFull for the real job, etc. If your 
>> Full’s ever screw up, you will want to do this (what I do for road warriors 
>> where Fulls are not really possible) to avoid the long time to do a new 
>> full from the source. Putting an Archive job back into a backup job, 
>> running a manual virtual full of a job will suck the files defined in the 
>> archive job in (won’t if it’s an archive, thus the change to update job 
>> type=B) and rebuild a full, then do an incremental. 
>>
>>
>> I personally do a version of case 2, but I use Migrate jobs to move the 
>> AI-Consolidated to a tape pool, but my jobs full’s are much smaller than 
>> yours. I also only do an offsite archive job once a month. Otherwise 
>> Incremental and AI-Consolidated on Disk, Migrate to tape pool purge 
>> AI-Consolidated jobs quickly. So when I consolidate it’s Tape + Disk —> 
>> Disk —> Migrate Tape —> Archive Tape 
>>
>> # copy job to long term tape 
>> Job { 
>> Name = "Migrate-To-Offsite-AI-Consolidated" 
>> Client = myth-fd 
>> Type = Migrate 
>> Purge Migration Job = yes 
>> Pool = AI-Consolidated 
>> Level = Full 
>> Next Pool = Offsite 
>> Schedule = WeeklyCycleAfterBackup 
>> Allow Duplicate Jobs = no 
>> Priority = 4 #before catalog dump 
>> Messages = Standard 
>> Selection Type = PoolTime # 7 days 
>> Spool Data = No 
>> Selection Pattern = "." 
>> RunAfterJob = "/usr/local/bin/prune.sh” 
>> } 
>>
>> The script forces the volumes to truncate/prune after I move the jobs 
>> off. 
>>
>>
>> Brock Palen 
>> 1 (989) 277-6075 <(989)%20277-6075> 
>> bro...@mlds-networks.com 
>> www.mlds-networks.com 
>> Websites, Linux, Hosting, Joomla, Consulting 
>>
>>
>>
>> > On Aug 5, 2020, at 5:40 AM, a.br...@mail.de <a.br...@mail.de> wrote: 
>> > 
>> > 
>> > Hello everybody, 
>> > 
>> > I'm in the process of developing a regular backup-strategy and found 
>> that I need some assistance. Here are the parameters in short: 
>> > - 35TB of medical imaging data 
>> > - daily increment of 50-60GB 
>> > - one site, 10Gb/s Backbone 
>> > - Overland NEOs LTO7 Storageloader, 8-bay. 
>> > attached to 
>> > - dedicated backupserver with 20TB RAID6, will be enhanced as needed 
>> > 
>> > I have already backed up all data on LTO7 tape (WORM, for legal 
>> reasons) as of Dec'19. 
>> > A 2nd generation, also LTO7 WORM, is currently in progress (VERY! slow, 
>> ~12MB/s, different story). Tapes are/will be stored offsite. 
>> > After that I'm planning to do a 3rd generation on disk inhouse and 
>> amend the 1st and 2nd gen on tape so that I end up with three generations 
>> identical to a certain cutoff date. 
>> > 
>> > Then, what to do next? How could a daily routine look like? 
>> > The radiologists are very concerned about their data and would like to 
>> see a daily tape change with the ejected tape being taken offsite in the 
>> evening. A 1-bay LTO8 drive could be purchased then, the daily tape change 
>> would be done by an apprentice or so... 
>> > 
>> > So I thought about an Always-Incremental-B2D2T-strategy starting with 
>> the above cutoff day. But I still have too less experience and so I'm 
>> struggeling hard to develop a clear structure in my head - WHAT is WHEN 
>> being copied WHERE - let alone transform that into a working bareos 
>> configuration. 
>> > 
>> > Do my thoughts appear reasonable up to that point? 
>> > BTW: can a daily tape change be realized at all, where you just push 
>> the button to eject the TUE-tape, insert the WED-tape, and so on without 
>> having to stop the storage-daemon in order to unlock the tape? 
>> > 
>> > Thanks for helping me with the first steps to bring this under way. 
>> > 
>> > Andreas 
>> > 
>> > 
>> > 
>> > -- 
>> > You received this message because you are subscribed to the Google 
>> Groups "bareos-users" group. 
>> > To unsubscribe from this group and stop receiving emails from it, send 
>> an email to bareos-users...@googlegroups.com. 
>> > To view this discussion on the web visit 
>> https://groups.google.com/d/msgid/bareos-users/64fe4cfa-6c87-4d28-aed9-9e30de285ee1n%40googlegroups.com.
>>  
>>
>>
>> -- 
> You received this message because you are subscribed to the Google Groups 
> "bareos-users" group.
> To unsubscribe from this group and stop receiving emails from it, send an 
> email to bareos-users...@googlegroups.com.
>
> To view this discussion on the web visit 
> https://groups.google.com/d/msgid/bareos-users/4aa25985-fd74-4e2b-bb71-95ed37154bc1n%40googlegroups.com
>  
> <https://groups.google.com/d/msgid/bareos-users/4aa25985-fd74-4e2b-bb71-95ed37154bc1n%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to bareos-users+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/bareos-users/414d9ef4-b729-43fb-abb2-1cd4cc5a2a7cn%40googlegroups.com.

Reply via email to