Hi Kern & Bacula-developers,

I have been working on changes to Bacula to enable chunked backup for large 
files,
such as multi-gigabyte virtual disks [which I have], and possibly database 
files etc.

I need to establish how and when the per-chunk hash values are retrieved from 
the
database and stored/updated to the database.

I am starting with backup changes, for obvious reasons, and note that the data 
stream
from FD -> SD is a single contiguous stream, albeit transferred in record sized 
pieces.

I was envisaging alternate data-chunk / chunk-hash transfers, but that does not 
fit as easily
into the existing code as I had hoped [src/stored/append.c and 
src/filed/backup.c].

Does the per-chunk hash value info also need to go onto the storage media as to 
the database ?

If it does, then I could simply accumulate the file-offset/hash-value pairs and 
send them as a separate
stream after the data, although that may be less than ideal in memory 
consumption terms.

For restore, the current code is configured such that the SD is unaware of the 
file-offset values for a
sparse data stream, which means that the SD would be unable to be selective 
about the data which
it sends to the FD, which is somewhat link-inefficient.

Any comments ?

Will you [Kern] be at the Amsterdam meeting at all ?

Regards,

Howard

--
"Only two things are infinite, the universe and human stupidity,
and I'm not sure about the former." -- Albert Einstein 


------------------------------------------------------------------------------
This SF.net email is sponsored by Sprint
What will you do first with EVO, the first 4G phone?
Visit sprint.com/first -- http://p.sf.net/sfu/sprint-com-first
_______________________________________________
Bacula-devel mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/bacula-devel

Reply via email to