ID: 33070
User updated by: lindsay at bitleap dot com
Reported By: lindsay at bitleap dot com
-Status: Feedback
+Status: Open
Bug Type: Performance problem
Operating System: Linux 2.6.10 kernel
PHP Version: 5.0.3
New Comment:
Given that bzip can be used in a stream, could the data be decompressed
in chunks?
Previous Comments:
------------------------------------------------------------------------
[2005-05-22 15:20:22] [EMAIL PROTECTED]
And how do you think it should work?
I don't see any more effective way that will not hit memory instead of
CPU usage.
------------------------------------------------------------------------
[2005-05-20 16:56:48] lindsay at bitleap dot com
Script:
<?php
decrypt( "256K", file_get_contents("256K.bz2", "r"));
decrypt( "512K", file_get_contents("512K.bz2", "r"));
decrypt( "1M", file_get_contents("1M.bz2", "r"));
decrypt( "2M", file_get_contents("2M.bz2", "r"));
decrypt( "3M", file_get_contents("3M.bz2", "r"));
decrypt( "4M", file_get_contents("4M.bz2", "r"));
function decrypt($file_name, $file_data)
{
echo "file length $file_name ran in: ";
$time_start = time();
bzdecompress($file_data, false);
$end_start = time();
echo ($end_start - $time_start) . " seconds\n";
}
?>
Data:
If you run linux:
dd if=/dev/urandom bs=1024 count=256 of=256K
dd if=/dev/urandom bs=1024 count=512 of=512K
dd if=/dev/urandom bs=1024 count=1024 of=1M
dd if=/dev/urandom bs=1024 count=2048 of=2M
dd if=/dev/urandom bs=1024 count=3072 of=3M
dd if=/dev/urandom bs=1024 count=4096 of=4M
bzip2 256K 512K 1M 2M 3M 4M
If not, let me know and I'll upload or email data samples.
------------------------------------------------------------------------
[2005-05-19 23:44:05] [EMAIL PROTECTED]
Please provide a short but complete reproduce script and (if possible)
the data too.
------------------------------------------------------------------------
[2005-05-19 17:22:47] lindsay at bitleap dot com
Description:
------------
I found bug #13860 regarding bzdecompress speeds on 4.2.x. The bug
suggests its fixed for 4.2.0 but I see slowness in 5.0.3.
On 5.0.3, bzdecompress seems to get exponentionally slower as data
sizes increase. I timed the length of decompression on increasing file
sizes containing compressed random data:
file length 256K ran in: 2 seconds
file length 512K ran in: 10 seconds
file length 1M ran in: 41 seconds
file length 2M ran in: 135 seconds
file length 3M ran in: 278 seconds
file length 4M ran in: 476 seconds
I'm not a c coder but the do while loop at line 472 of:
http://cvs.php.net/co.php/php-src/ext/bz2/bz2.c?r=1.9
seems to decompress the string over and over. Each iteration attempts
a decompress w/ a buffer size based on the iteration count. If the
decompress fails, a larger buffer is tried. If I'm reading this
correctly, the same data could be decompressed hundreds of times until
the buffer gets large enough.
------------------------------------------------------------------------
--
Edit this bug report at http://bugs.php.net/?id=33070&edit=1