Thanks guys- I think I'll have to do multiple queries using LIMIT as Geoff
suggested since apparently mysql_unbuffered_query() would lose the result
set of the "select * from" query once I run the insert query.  I'm still not
sure why the MYSQL_CLIENT_COMPRESS didn't seem to have an effect, however.

David

-----Original Message-----
From: Chris [mailto:[EMAIL PROTECTED]
Sent: Monday, February 06, 2006 4:16 PM
To: David Yee
Cc: 'php-general@lists.php.net'
Subject: Re: [PHP] PHP mysql data result set compression


Hi David,

See http://www.php.net/mysql_unbuffered_query

It won't load the whole lot into memory before returning it to php.

David Yee wrote:
> Hi all- is there a way have a large data result set from MySQL compressed?
> E.g. I have a table with over a million rows of data that I want to do a
> "select * from " on and then take that result, do some field/data
> manpulation, and then insert row-by-row to another table.  The problem is
> the result of the query is so big that it's casuing PHP to swap to disk,
> causing things to slow to a crawl.  Doing a "show processlist" on the
mysql
> console shows that "Writing to net" is the state of the running "select *
> from " query.  I tried adding the flag "MYSQL_CLIENT_COMPRESS" to both
> mysql_pconnect() and mysql_connect() but it doesn't seem to do any
> compression (I can tell by the size of the running php memory process).
Any
> ideas would be appreciated- thanks.
> 
> David
> 

-- 
PHP General Mailing List (http://www.php.net/)
To unsubscribe, visit: http://www.php.net/unsub.php

Reply via email to