I was curious if lftp or wget will ever support a rollback feature and
somehow verify the bytes are correct somehow where the file has been
resumed.
Why? This is stated below:
LFTP VS WGET EXPERIMENT:
PROBLEM: With lftp, many of my downloads get corrupted.
This is because the connection between my satellite link, and
my ISP
gets severed, therefore causing FTP to resume. Regular
connection
breaks are normal on a satellite connection, they may only last
1ms or
less, however, they cause the FTP to resume, thus causing
corruption.
QUESTION: However, does lftp corrupt files more often than say wget?
TEST LFTP: With each 700MB pull with lftp, 6% of files usually are bad.
This means 3 to 4 re-downloads. I've downloaded over 1
terrabyte
of data, the average seems to be about 6%, the more resumes,
the
greater the percentage file problems.
TEST WGET: First 700MB file transfer: 0.00% corruption.
Second 700MB file transfer: 0.00% corruption.
Third 700MB file transfer: 5.00% corruption.
After several more 700MB pulls, it is about the same.
POINT: It is a single character error. I've done multiple splits and
diffs.
I've found only 1 character is different from the original.
This is however, catastrophic for binary files.