Hi Patrick,

i did a short search in our documentation how we did this.
If you can start from scratch you can do it like that.
It is required that you have splited your backup jobs before.

1. Disabling all jobs
2. Run job or a bunch of jobs on different days.

To get that automatically we had used that python code i hope it will help you.

----------
 cmd = 'echo "run"|pfexec bconsole |awk \'/[0-9]:/ && /\./ {print $2}\' |grep 
-e "p.oew.de"'
#cmd = 'echo "run"|pfexec bconsole |awk \'/[0-9]:/ && /\./ {print $2}\''
#cmd = 'echo "run"|sudo bconsole |awk \'/[0-9]:/ && /\./ {print $2}\''
bacula_max_full_interval = 30
job_start_time = '18:15:10'
 
from subprocess import *
import re
import datetime
 
p = Popen(cmd, shell=True, stdin=PIPE, stdout=PIPE, stderr=STDOUT, 
close_fds=True)
joblist_global = p.stdout.read()
 
joblist = []
for job in joblist_global.splitlines():
 joblist.append(job)
 
for i in xrange(bacula_max_full_interval):
 # http://docs.python.org/release/2.3/whatsnew/section-slices.html
 # create a number of 'bacula_max_full_interval' partitions of our joblist, 
starting with item 'i'
 job_start_day = (datetime.date.today() + datetime.timedelta(i)).isoformat() 
 for j in joblist[i::bacula_max_full_interval]:
  print 'disable job=' + j
  #print 'run job=' + j + ' level=Full when="' + job_start_day + ' ' + 
job_start_time + '" yes'
 
print
for i in xrange(bacula_max_full_interval):
 # http://docs.python.org/release/2.3/whatsnew/section-slices.html
 # create a number of 'bacula_max_full_interval' partitions of our joblist, 
starting with item 'i'
 job_start_day = (datetime.date.today() + datetime.timedelta(i)).isoformat() 
 for j in joblist[i::bacula_max_full_interval]:
  print 'run job=' + j + ' when="' + job_start_day + ' ' + job_start_time + '" 
yes'
----------

If you can't start from scratch you could split out a part of data into a new 
job and exclude it in the old job. After a couple of days you will have all 
data from one job in, lets say, 100 jobs.
These 100 jobs you can distribute over ... 60 days and with max full interval 
you can manage when the next virtual full is done...

Regards Marco

On Wednesday, October 21, 2015 at 1:01:14 PM UTC+2, Patrick Glomski wrote:
> Thank you, Marco! I will try splitting the backup of the large system into 
> several smaller jobs and staggering it. If that is feasible with the 
> number/size of files in the production storage, it sounds like an excellent 
> solution!
> 
> Patrick

-- 
You received this message because you are subscribed to the Google Groups 
"bareos-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
For more options, visit https://groups.google.com/d/optout.

Reply via email to