Hi win.a,
I tried with another user the problem still persists. I am getting same
error for another user as well.
--
Regards,
Manasi Save
On Tue, 7 Sep 2010 11:41:35 0800, win.a wrote:
use another user and dump the data ,eg the root .
mysqldump -uroot -p --al-databases
: Wed, September 8, 2010 3:57:40 PM
Subject: Re: mysqldump error 1064 for database Use command
Hi win.a,
I tried with another user the problem still persists. I am getting same
error for another user as well.
--
Regards,
Manasi Save
On Tue, 7 Sep 2010 11:41:35 0800, win.a wrote:
use another
On 06/09/2010 6:54 a, Manasi Save wrote:
Dear Nitin,
I have newly installed mysql on this server.
mysql Select version();
+-+
| version() |
+-+
| 5.1.22-rc-Debian_2~ppa5-log |
+-+
Earlier I use to run the same
No it does not. But when I dump database name mydb it does. but not
the database with name 1.
--
Regards,
Manasi Save
On Mon, 06 Sep 2010 09:23:21 0200, Jangita wrote:
On 06/09/2010 6:54 a, Manasi Save wrote:
Dear Nitin,
I have newly installed mysql on this server.
mysql
Dear Nitin,
I have newly installed mysql on this server.
mysql Select version();
+-+
| version() |
+-+
| 5.1.22-rc-Debian_2~ppa5-log |
+-+
Earlier I use to run the same command on Fedora-with same mysql
Hi,
Have you recently upgraded your MySQL installation? 1064 in earlier days used
to
mean use of reserved word. Few clients have reported this error in last couple
of months after they upgraded from 5.1.41 to higher versions. As of now, I
don't
have a solution other than changing the name
MySQL dump calls LOCK TABLES before dumping the data (presumably to
prevent data modification halfway through the process).
LOCK TABLES has its own privilege (conveniently given the same name)
which your user account will have to be given in order to run a
mysqldump.
On Fri, 2009-05-01 at 15:49
On Tue, 2009-01-13 at 12:19 +0530, Chandru wrote:
Hi,
Did u try using this command
mysqldump --opt db_name db_name.sql -p 2bkp.err
Not quite. Firstly, I had to alter the normal backup cron job, and that
doesn't happen until late at night.
Secondly, yes I added the redirection to
This sounds like you need to raise max_allowed_packet for mysqldump
(and possibly mysqld) - these are separate settings for both the
client and the server. You can do this via the my.cnf (or ~/.my.cnf)
or specify it as an option on the command line mysqldump --opt ...
--max_allowed_packet=1G
On Tue, 13 Jan 2009 18:34:44 -0600, Andrew Garner
andrew.b.gar...@gmail.com wrote:
This sounds like you need to raise max_allowed_packet for mysqldump
(and possibly mysqld) - these are separate settings for both the
client and the server. You can do this via the my.cnf (or ~/.my.cnf)
or
On Tue, Jan 13, 2009 at 6:06 PM, Dan d...@entropy.homelinux.org wrote:
On Tue, 13 Jan 2009 18:34:44 -0600, Andrew Garner
andrew.b.gar...@gmail.com wrote:
This sounds like you need to raise max_allowed_packet for mysqldump
(and possibly mysqld) - these are separate settings for both the
Hi,
please increase your interactive_timeout variable to some big number and
also try to log the erros if any thing by using the command:
mysqldump --opt db_name db_name.sql -p 2bkp.err
check if you get some thing in the bkp.err file.
Regards,
Chandru,
www.mafiree.com
On Mon, Jan 12, 2009
On Mon, 12 Jan 2009 16:25:12 +0530, Chandru chandru@gmail.com wrote:
Hi,
please increase your interactive_timeout variable to some big number and
also try to log the erros if any thing by using the command:
mysqldump --opt db_name db_name.sql -p 2bkp.err
check if you get some
I'm also having a similar issue with some tables I've been trying to dump
(total data set is around 3TB). I'm dumping directly from one host to
another (mysqldump -hSOURCE DATABASE | mysql -hLOCALHOST DATABASE) using
mysql 4.1.22. One system is Solaris 10 SPARC, while the other is Solaris 10
x64
Show the details of your hardware us.
On Thu, Oct 2, 2008 at 3:02 PM, Krishna Chandra Prajapati
[EMAIL PROTECTED] wrote:
Hi,
Just try the below command on console. It will give that the error is
exactly related to what.
$perror 5
What is total ram in your box.
On Thu, Oct 2, 2008 at
Hi,
Just try the below command on console. It will give that the error is
exactly related to what.
$perror 5
What is total ram in your box.
On Thu, Oct 2, 2008 at 12:26 PM, Uma Bhat [EMAIL PROTECTED] wrote:
Hey Guys!
I have been googling a lot on this error and read various suggestions.
I am getting the below error
[EMAIL PROTECTED] drupal]# mysqldump --tab=/usr/local/drupal/drupalbckup/ --opt
drupal
mysqldump: Got error: 1: Can't create/write to file
'/usr/local/drupal/drupalbckup/access.txt' (Errcode: 13) when
executing 'SELECT INTO OUTFILE'
Change the owner of that
[EMAIL PROTECTED] drupal]# mysqldump --tab=/usr/local/drupal/drupalbckup/
--opt
drupal
mysqldump: Got error: 1: Can't create/write to file
'/usr/local/drupal/drupalbckup/access.txt' (Errcode: 13) when
executing 'SELECT INTO OUTFILE'
Change the owner of that directory so the mysql server
Hello everyone!
I just wanted to give everyone an update. I'm still getting this error
when I try to back up this database table. I don't get it at the same
row each time - today was at row 1,618, yesterday it was at row 24,566.
Just a reminder of my symptoms:
1. mysqldump is the only thing
Have a look here:
http://dev.mysql.com/doc/mysql/en/gone-away.html
Gleb,
Thanks for the response. The only one that seems to apply is this one:
You may also see the MySQL server has gone away error if
MySQL is started with the --skip-networking option.
I do start mySQL without networking
Gustafson, Tim [EMAIL PROTECTED] wrote on 08/29/2005 09:24:36 AM:
Hello
I am using mysqldump to backup my entire database (about 40GB total)
each night. I dump each table separetly, so that if mysqldump crashes
in the middle somewhere, the rest of the database still gets backed up.
Most
/
-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED]
Sent: Monday, August 29, 2005 9:35 AM
To: Gustafson, Tim
Cc: mysql@lists.mysql.com
Subject: Re: mysqldump: Error 2013
Gustafson, Tim [EMAIL PROTECTED] wrote on 08/29/2005 09:24:36 AM:
Hello
I am using
Gustafson, Tim wrote:
When backing up this table, I get this error message every night:
/usr/local/bin/mysqldump: Error 2013: Lost connection to MySQL server
during query when dumping table `DocumentVariants` at row: 13456
The table actually has 94,916 rows in it. There are no entries in the
: Hassan Schroeder [mailto:[EMAIL PROTECTED]
Sent: Monday, August 29, 2005 9:51 AM
To: mysql@lists.mysql.com
Subject: Re: mysqldump: Error 2013
Gustafson, Tim wrote:
When backing up this table, I get this error message every night:
/usr/local/bin/mysqldump: Error 2013: Lost connection
Gustafson, Tim wrote:
No, max_allowed_packet is 100 megabytes, and the maximum data field
in the database is 50MB right now, and most are well below 10MB.
mmm. OK, not having any more bright ideas :-) I would try dumping
it using an explicit `--where=` claus to guarantee fixed ordering,
and
Hello.
Has anyone else had similar problems? Can anyone shed some light on how
to successfully back up this database?
Have a look here:
http://dev.mysql.com/doc/mysql/en/gone-away.html
Hello
I am using mysqldump to backup my entire database (about 40GB total)
each night. I
Hassan Schroeder wrote:
Does the size of the contents of that field exceed your defined
max_allowed_packet size?
Gustafson, Tim wrote:
No, max_allowed_packet is 100 megabytes, and the maximum data field in
the database is 50MB right now, and most are well below 10MB.
Tim Gustafson
MEI
I believe it's the size of the row, not the size of a
single field, that matters. Is it possible you have a
row which exceeds max_allowed_packet size?
No. There is one blob fields (always less than 50MB) and like 10 other fields,
all integers.
smime.p7s
Description: S/MIME cryptographic
Hello.
Try '--open-files-limit=8192' at least. Check the real value of
open_file_limits with such statement:
show variables like '%open_f%';
You can find some recommendations for SuSe Linux at:
http://dev.mysql.com/doc/mysql/en/Linux-post-install.html
Mysql user [EMAIL PROTECTED]
Yep, that seems to have done it, at least when I'm testing it. I'm
pretty sure the problem is fixed, but the next couple of days of
automated backups will tell.
Thanks!
On Tue, 2005-01-11 at 00:43, Gleb Paharenko wrote:
Hello.
Try '--open-files-limit=8192' at least. Check the real value of
At 15:47 -0400 10/27/04, Michael Ragsdale wrote:
Running MySQL 4.0.12-NT on a Windows 2k Pro Server.
When I attempt to execute mysqldump.exe, I get the following error:
mysqldump: ERROR: unknown variable 'local-infile=1'
I tried using --force and it didn't help.
Showing the variables, I get
Hi.
See
http://dev.mysql.com/doc/mysql/en/LOAD_DATA_LOCAL.html
and upgrade to the latest stable release.
Michael Ragsdale [EMAIL PROTECTED] wrote:
Running MySQL 4.0.12-NT on a Windows 2k Pro Server.
When I attempt to execute mysqldump.exe, I get the following error:
First, mysqldump is a command in and of itself so it does NOT execute from
within the mysql command line tool. Next, expand your command entry a bit
to include your userid and password, for example:
mysqldump -uuserid -ppassword --add-drop-table mtweblog BckUpmtweblog.sql
-Original
-ppassword --no-data mtweblog mtweblog.ddl
mysql -uuser -ppassword mtweblog.ddl
mysql -uuser -ppassword bkupmtweblog.sql
-Original Message-
From: tait sanders [mailto:[EMAIL PROTECTED]
Sent: Sunday, March 21, 2004 5:13 PM
To: [EMAIL PROTECTED]
Cc: mysql
Subject: Re: mysqldump error
thanks - i ran the command below and it worked. and now I'm trying to
import using:
'mysql -u root -p mtweblog Bckupmtweblog.sql' on CLI on different
machine and am getting Error 1 at line 11: Can't create/write to
file'./mtweblog/mt_author.frm' Errorcode 13.
any suggestions here?
tait
On
At 10:12 +1100 3/22/04, tait sanders wrote:
thanks - i ran the command below and it worked. and now I'm trying
to import using:
'mysql -u root -p mtweblog Bckupmtweblog.sql' on CLI on different
machine and am getting Error 1 at line 11: Can't create/write to
file'./mtweblog/mt_author.frm'
Try putting the name inside of backticks;
From the Manual:
6.1.2 Database, Table, Index, Column, and Alias Names
---
Note that if the identifier is a restricted word or contains special characters you
must always
quote it with a
Rusty Wright [EMAIL PROTECTED] wrote:
I have a user who unfortunately named one of his tables order. He also
has table names with a dash in them. mysqldump is unable to dump the
order table and I'm unable to use the ALTER command on it and the tables
with the dashes in their names to rename
`TableNamesOr FieldNames`
- Original Message -
From: Kelley Lingerfelt [EMAIL PROTECTED]
To: Rusty Wright [EMAIL PROTECTED]; Mysql List
[EMAIL PROTECTED]
Sent: Monday, September 29, 2003 09:10
Subject: Re: mysqldump error (mysql 4.0.14)
Try putting the name inside of backticks;
From
- Original Message -
From: David Angela Ehmer [EMAIL PROTECTED]
To: [EMAIL PROTECTED]
Sent: Saturday, January 04, 2003 11:00 PM
Subject: mysqldump ERROR 1064
Hi
I am having problems generating the mysqldump file. I have tried a range
of
options and carefully studied several
On Sun, 2003-01-05 at 00:00, David Angela Ehmer wrote:
Hi
I am having problems generating the mysqldump file. I have tried a range of
options and carefully studied several tutorials I have on using this
command. Each time I get the following errror.
ERROR 1064 you have an error in your
mysql 3.23.49a
From: Stefan Hinz, iConnect \(Berlin\) [EMAIL PROTECTED]
Reply-To: Stefan Hinz, iConnect \(Berlin\) [EMAIL PROTECTED]
To: Carlos A. [EMAIL PROTECTED]
CC: [EMAIL PROTECTED]
Subject: Re: mysqldump error
Date: Thu, 5 Dec 2002 22:02:49 +0100
Dear Carlos,
ERROR 2013 at line 1
Dear Carlos,
Got error: 2013: Lost connection to MySQL server
during query
when retrieving data from server
In your my.cnf / my.ini configuration file, set the value for the
communication buffer high (maximum is 16 MB):
[mysqldump]
set-variable = max_allowed_packet=16M
Restart the server,
think is problem with the binary, but I don't know why.. an idea may be
install a recent new version from mysql.
From: Listen Hinz [EMAIL PROTECTED]
To: Carlos A. [EMAIL PROTECTED]
CC: [EMAIL PROTECTED]
Subject: Re: mysqldump error
Date: Thu, 5 Dec 2002 18:16:40 +0100
Dear Carlos,
Got
iConnect GmbH http://iConnect.de
Heesestr. 6, 12169 Berlin (Germany)
Telefon: +49 30 7970948-0 Fax: +49 30 7970948-3
- Original Message -
From: Carlos A. [EMAIL PROTECTED]
To: [EMAIL PROTECTED]
Cc: [EMAIL PROTECTED]
Sent: Thursday, December 05, 2002 7:25 PM
Subject: Re: mysqldump error
remotely there is no problem, the problem is
local. I think is a version problem.
From: Stefan Hinz, iConnect \(Berlin\) [EMAIL PROTECTED]
Reply-To: Stefan Hinz, iConnect \(Berlin\) [EMAIL PROTECTED]
To: Carlos A. [EMAIL PROTECTED]
CC: [EMAIL PROTECTED]
Subject: Re: mysqldump error
Date: Thu
tl,
Thursday, September 26, 2002, 3:44:30 PM, you wrote:
t Please help me mysqldump:
t ON:
t mysqldump --opt --user=user_name -pPassword _stat_ | gzip -1 database.gz
t I have an error :
t mysqldump: Got error: 1017: Can't find file: './_stat_/_stat_193_219_2_250.frm'
(errno: 24) when using
-
From: "Gerald L. Clark" [EMAIL PROTECTED]
To: "Joel Caturia" [EMAIL PROTECTED]
Cc: [EMAIL PROTECTED]
Sent: Wednesday, March 14, 2001 1:43 PM
Subject: Re: mysqldump Error
Joel Caturia wrote:
I'm trying to use mysqldump to dump a database (obviously).
Here's the comma
Hi,
Error 13 is permissions. Your server has to have permission to write to the
directory, and over-write any existing file of the same name.
Regards
Quentin
-Original Message-
From: Joel Caturia [mailto:[EMAIL PROTECTED]]
Sent: Thursday, 15 March 2001 10:07 a.m.
To: [EMAIL PROTECTED]
Hi,
But is your mysql server run by root, or, as advised, by 'mysql'?
Quentin
-Original Message-
From: Joel Caturia [mailto:[EMAIL PROTECTED]]
Sent: Thursday, 15 March 2001 10:56 a.m.
To: Gerald L. Clark
Cc: [EMAIL PROTECTED]
Subject: Re: mysqldump Error
ah but I do.. Allow me
-
From: "Quentin Bennett" [EMAIL PROTECTED]
To: "'Joel Caturia'" [EMAIL PROTECTED]; "Gerald L. Clark"
[EMAIL PROTECTED]
Cc: [EMAIL PROTECTED]
Sent: Wednesday, March 14, 2001 2:19 PM
Subject: RE: mysqldump Error
Hi,
But is your mysql server run by roo
51 matches
Mail list logo