This sounds like you need to raise max_allowed_packet for mysqldump
(and possibly mysqld) - these are separate settings for both the
client and the server. You can do this via the my.cnf (or ~/.my.cnf)
or specify it as an option on the command line mysqldump --opt ...
--max_allowed_packet=1G
On Tue, 13 Jan 2009 18:34:44 -0600, Andrew Garner
andrew.b.gar...@gmail.com wrote:
This sounds like you need to raise max_allowed_packet for mysqldump
(and possibly mysqld) - these are separate settings for both the
client and the server. You can do this via the my.cnf (or ~/.my.cnf)
or
On Tue, Jan 13, 2009 at 6:06 PM, Dan d...@entropy.homelinux.org wrote:
On Tue, 13 Jan 2009 18:34:44 -0600, Andrew Garner
andrew.b.gar...@gmail.com wrote:
This sounds like you need to raise max_allowed_packet for mysqldump
(and possibly mysqld) - these are separate settings for both the
Hi all. I have a 30GB innodb-only database in mysql-5.0.54. I have
always done nightly backups with:
mysqldump --opt db_name db_name.sql -p
Recently this started failing with:
Error 2013: Lost connection to MySQL server
I have checked all tables for corruption - nothing found. Also as far as
I
Hi,
please increase your interactive_timeout variable to some big number and
also try to log the erros if any thing by using the command:
mysqldump --opt db_name db_name.sql -p 2bkp.err
check if you get some thing in the bkp.err file.
Regards,
Chandru,
www.mafiree.com
On Mon, Jan 12, 2009
On Mon, 12 Jan 2009 16:25:12 +0530, Chandru chandru@gmail.com wrote:
Hi,
please increase your interactive_timeout variable to some big number and
also try to log the erros if any thing by using the command:
mysqldump --opt db_name db_name.sql -p 2bkp.err
check if you get some
I'm also having a similar issue with some tables I've been trying to dump
(total data set is around 3TB). I'm dumping directly from one host to
another (mysqldump -hSOURCE DATABASE | mysql -hLOCALHOST DATABASE) using
mysql 4.1.22. One system is Solaris 10 SPARC, while the other is Solaris 10
x64