Laurence Perkins wrote:
>>
>> Curious question here.  As you may recall, I backup to a external hard 
>> drive.  Would it make sense to use that software for a external hard drive?  
>> Right now, I'm just doing file updates with rsync and the drive is 
>> encrypted.  Thing is, I'm going to have to split into three drives soon.  
>> So, compressing may help.  Since it is video files, it may not help much but 
>> I'm not sure about that.  Just curious. 
>>
>> Dale
>>
>> :-)  :-) 
>>
>>
> If I understand correctly you're using rsync+tar and then keeping a set of 
> copies of various ages.

Actually, it is uncompressed and just stores one version and one copy. 


>
> If you lose a single file that you want to restore and have to go hunting for 
> it, with tar you can only list the files in the archive by reading the entire 
> thing into memory and only extract by reading from the beginning until you 
> stumble across the matching filename.  So with large archives to hunt 
> through, that could take...  a while...
>
> dar is compatible with tar (Pretty sure, would have to look again, but I 
> remember that being one of its main selling-points) but adds an index at the 
> end of the file allowing listing of the contents and jumping to particular 
> files without having to read the entire thing.  Won't help with your space 
> shortage, but will make searching and single-file restores much faster.
>
> Duplicity and similar has the indices, and additionally a full+incremental 
> scheme.  So searching is reasonably quick, and restoring likewise doesn't 
> have to grovel over all the data.  It can be slower than tar or dar for 
> restore though because it has to restore first from the full, and then walk 
> through however many incrementals are necessary to get the version you want.  
> This comes with a substantial space savings though as each set of archive 
> files after the full contains only the pieces which actually changed.  
> Coupled with compression, that might solve your space issues for a while 
> longer.
>
> Borg and similar break the files into variable-size chunks and store each 
> chunk indexed by its content hash.  So each chunk gets stored exactly once 
> regardless of how many times it may occur in the data set.  Backups then 
> become simply lists of file attributes and what chunks they contain.  This 
> results both in storing only changes between backup runs and in deduplication 
> of commonly-occurring data chunks across the entire backup.  The 
> database-like structure also means that all backups can be searched and 
> restored from in roughly equal amounts of time and that backup sets can be 
> deleted in any order.  Many of them (Borg included) also allow mounting 
> backup sets via FUSE.  The disadvantage is that restore requires a compatible 
> version of the backup tool rather than just a generic utility.
>
> LMP


I guess that is the downside of not having just plain uncompressed
files.  Thing is, so far, I've never needed to restore a single file or
even several files.  So it's not a big deal for me.  If I accidentally
delete something tho, that could be a problem, if it has left the trash
already. 

Since the drive also uses LVM, someone mentioned using snapshots.  Still
not real clear on those even tho I've read a bit about them.  Some of
the backup technics are confusing to me.  I get plain files, even
incremental to a extent but some of the new stuff just muddies the water. 

I really need to just build a file server, RAID or something.  :/

Dale

:-)  :-)

Reply via email to