> This is a terrible assumption in general (i.e. if filesize % blocksize 
> is close to uniformly distributed).  If you remove one byte and the data 
> is stored with blocksize B, then you either save zero bytes with 
> probability 1-1/B or you save B bytes with probability 1/B.  The 
> expected number of bytes saved is B*1/B=1.  Since expectation is linear, 
> if you remove x bytes, the expected number of bytes saved is x (even if 
> there is more than one byte removed per file).

You didn't calculate the probability of actually saving a full block 
or not (that's the only thing that matters). I assumed it's relatively
small and can be ignored in practice since the amount of end white
space is negligible compared to total file size.

-Andi
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to