ID:               40494
 User updated by:  foster dot graeme at gmail dot com
 Reported By:      foster dot graeme at gmail dot com
 Status:           Bogus
 Bug Type:         Zip Related
 Operating System: Linux
 PHP Version:      5.2.1
 Assigned To:      pajoye
 New Comment:

Would it be possible to add a brief description of this situation to
the documentation, for example the following could be added to the
description of ZipArchive::addFile

Description

bool ZipArchive::addFile ( string filename [, string localname] )

Adds a link to the ZIP archive from a given path. When the archive is
closed the link is checked to ensure that the file still exists and
will then be compressed and added to the archive. If a lot of files are
being added then the number of file handles permitted by the OS may be
exceeded, if that occurs then the status will be set to
ZIPARCHIVE::ER_OPEN. This can be avoided by closing the archive before
the limit is reached and then reopening the archive.

for example:

if ($zip->numfile % $limit == 0)
{
   $zip->close();
   $zip->open($filename,ZIPARCHIVE::CREATE);
}


Previous Comments:
------------------------------------------------------------------------

[2007-02-15 14:35:34] [EMAIL PROTECTED]

"I still think that it would be nice if there was some way for the
system to manage this."

It is in the TODO list. As I said three times already in this
discussion. The solution is to add different modes:
- commit at the end when the archive is close
- immediate addition (will be much slower)

And again, it is in my TODOs already. I cannot tell when they will be
available (I do it on my free time).

In the meantime a simple:

if (($zip->numFiles % $yourlimit) == 0) {close; reopen;}

will do it.



"the archive can be partially built prior to the ulimit being reached.
This could be set as 250, with the ability to overload it. Maybe this
would only be triggered if a flag was set when the archive was
opened."

This solution does not work.The limit is arbitrary. There is no way to
get an exact value (and I doubt php is the only running process).


------------------------------------------------------------------------

[2007-02-15 14:02:51] foster dot graeme at gmail dot com

Okay thanks for the explanation, I understand the problem a little
better. I still think that it would be nice if there was some way for
the system to manage this.

I was thinking along the lines of a function to flush the files so that
the archive can be partially built prior to the ulimit being reached.
This could be set as 250, with the ability to overload it. Maybe this
would only be triggered if a flag was set when the archive was opened.

------------------------------------------------------------------------

[2007-02-15 13:23:36] [EMAIL PROTECTED]

See:

http://pecl.php.net/bugs/bug.php?id=9443

"it would be good if this wasn't necessary, in thatthe code could catch
the problem and allocate extra file handles if that is the problem."

This is not something I can control. The operating system defines it
and there is no way for me to increase this value.

I suggest you to close and reopen it every 1000 or so (or even 255 if
you want to go on the safest way, ie old windows).

Future releases will have a different mode, where the checks will done
only when you close the archives.

------------------------------------------------------------------------

[2007-02-15 13:14:57] foster dot graeme at gmail dot com

Maybe I need to explain this problem a little more.

I am trying to archive a folder on the server, at the moment it
contains 5609 folders and 11,221 files. The script loops through the
files adding them to the archive using the addFile() method. After the
first 1002 files I get a ZIPARCHIVE::ER_OPEN. If I close the archive
and the open it again I still have that error. However, if I close the
archive and open it before I get that error then I can archive all
11,221 files.

Since closing the file and re-opening fixes the problem (so long as I
do that before I get the error) Then may I suggest that closing an
archive will clear the status. Obviously, it would be good if this
wasn't necessary, in thatthe code could catch the problem and allocate
extra file handles if that is the problem.

------------------------------------------------------------------------

[2007-02-15 11:41:24] [EMAIL PROTECTED]

"When adding files to an archive, (using successive
ZipArchive::addFile()
commands) the compression doesn't happen until the file is closed. "

Yes, we do it while finalizing the archive.

" This can result in an out of memory error, "

You will run out of file ID before running out of memory. It  does not
really use many memory, only the file names and file handlers.

I suppose you are talking about the file handlers? 

"It would certainly require a rewrite of the ugly function
zip_close()"

What is ugly in this function? Or do you have a portable way to lock a
file until the archive creation is done?

I think you refer to the file handlers limitation. There is already a
bug about it and I plan to add a special (less safe) mode. This mode
will allow one to add only the paths without checks, errors will occur
only when the archive is closed. But that's a feature addition not a
bug fix.

I close this bug (not a bug > bogus).

Thanks for your report!

------------------------------------------------------------------------

The remainder of the comments for this report are too long. To view
the rest of the comments, please view the bug report online at
    http://bugs.php.net/40494

-- 
Edit this bug report at http://bugs.php.net/?id=40494&edit=1

Reply via email to