Hi Kern & Bacula-developers, I have been working on changes to Bacula to enable chunked backup for large files, such as multi-gigabyte virtual disks [which I have], and possibly database files etc.
I need to establish how and when the per-chunk hash values are retrieved from the database and stored/updated to the database. I am starting with backup changes, for obvious reasons, and note that the data stream from FD -> SD is a single contiguous stream, albeit transferred in record sized pieces. I was envisaging alternate data-chunk / chunk-hash transfers, but that does not fit as easily into the existing code as I had hoped [src/stored/append.c and src/filed/backup.c]. Does the per-chunk hash value info also need to go onto the storage media as to the database ? If it does, then I could simply accumulate the file-offset/hash-value pairs and send them as a separate stream after the data, although that may be less than ideal in memory consumption terms. For restore, the current code is configured such that the SD is unaware of the file-offset values for a sparse data stream, which means that the SD would be unable to be selective about the data which it sends to the FD, which is somewhat link-inefficient. Any comments ? Will you [Kern] be at the Amsterdam meeting at all ? Regards, Howard -- "Only two things are infinite, the universe and human stupidity, and I'm not sure about the former." -- Albert Einstein ------------------------------------------------------------------------------ This SF.net email is sponsored by Sprint What will you do first with EVO, the first 4G phone? Visit sprint.com/first -- http://p.sf.net/sfu/sprint-com-first _______________________________________________ Bacula-devel mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/bacula-devel
