On Tue, July 17, 2007 13:31, Baron Schwartz wrote:
> Mogens Melander wrote:
>> On Tue, July 17, 2007 04:29, Baron Schwartz wrote:
>>> B. Keith Murphy wrote:
>>>>
>>>> The problem is that I am realizing that this dump/import is going to
>>>> take
>>>> hours and in some cases days. I am looking for any way to speed this up.
>>>> Any suggestions?
>>>
>>> The fastest way I've found is to do SELECT INTO OUTFILE on the master,
>>> which
>>> selects into a sort of tab-delimited format by default -- don't specify
>>> any
>>> options like field terminators or whatnot.  This file can then be imported
>>> directly into LOAD DATA INFILE, again without options.
>>>
>>> I think this is faster than loading files full of SQL statements, which
>>> have to be parsed and query-planned etc.

That method has proven "very" quick in the past.

>>>
>>> I thought mysqldump had an option to dump this way, but I can't see it
>>> now.
>>
>> I think you are looking for the --single-transaction option :)
>
> I found the option I meant:
>
>    -T, --tab=name      Creates tab separated textfile for each table to given
>                        path. (creates .sql and .txt files). NOTE: This only
>                        works if mysqldump is run on the same machine as the
>                        mysqld daemon.
>

Yup, that was what i was trying to write 8^) using this one with the other.


-- 
Later

Mogens Melander
+45 40 85 71 38
+66 870 133 224



-- 
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.


-- 
MySQL General Mailing List
For list archives: http://lists.mysql.com/mysql
To unsubscribe:    http://lists.mysql.com/[EMAIL PROTECTED]

Reply via email to