** Description changed: Hello - In my case, I extract large archives (~1.5Gb in compressed state 7zip --> ~5Gb in uncompressed state). Extracted copy != oroginal, I compare files with vbindiff. + In my case, I extract large archives (~1.5Gb in compressed state 7zip --> ~5Gb in uncompressed state). Extracted copy != original, I compare files with vbindiff. I do not know is whether this information because in each extract data is lost in various areas. Example (in extracted copy): 199B CD5D - 199B CFFF Filled by 0x00h 2E79 A438 - 2E79 AFFF Filled by 0x00h 36B9 93DB - 36B9 9FFF Filled by 0x00h 71EF 1E45 - 71EF 1FFF Filled by 0x00h 7AFF 0C5A - 7AFF 0FFF Filled by 0x00h 95CE D752 - 95CF DFFF Filled by 0x00h ... etc ... As can be seen always missing data are replaced by zeros and block sizes are different, but justified on a boundary x0FFF. Ubuntu 11.04 Natty x86 2.6.37-9-generic-pae Filesystem: Ext4
-- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/691454 Title: Data is lost when writing large amounts of data to disk -- ubuntu-bugs mailing list ubuntu-bugs@lists.ubuntu.com https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs