hello,
assume the following table:
CREATE TABLE t (
id INT UNSIGNED auto_increment PRIMARY KEY,
c1 INT UNSIGNED NOT NULL,
c2 INT UNSIGNED NOT NULL,
c3 INT UNSIGNED,
UNIQUE (c1, c2, c3)
) engine = InnoDB;
Our first issue is that the UNIQUE constraint on (c1,c2,c3) does not
work in the case
: Daniel Kasak [mailto:[EMAIL PROTECTED]
Sent: Tuesday, April 18, 2006 11:35 PM
To: Tim Lucia
Cc: mysql@lists.mysql.com
Subject: Re: Document / Image (Blob) archival -- Best Practices
Tim Lucia wrote:
Hi all,
I am considering moving some archival records which largely consist of
blobs
(PDF Image
Hi all,
I am considering moving some archival records which largely consist of blobs
(PDF Image files) out of an Oracle DB and onto MySQL. Has anyone done this
(not necessarily the Oracle part) who can relate their experience(s)? I can
go with MyISAM or archive storage engine, from the looks
Tim,
I did this for a large collection of images, ~1million images up
around 40 gigs total last time I checked (no longer involved in the
project). It worked very well, performance was not horrible compared
to file-based storage. I always feared that MyISAM table getting
corrupted and
Tim Lucia wrote:
Hi all,
I am considering moving some archival records which largely consist of blobs
(PDF Image files) out of an Oracle DB and onto MySQL. Has anyone done this
(not necessarily the Oracle part) who can relate their experience(s)? I can
go with MyISAM or archive storage
Answers intermingled below
--- Bruno B B Magalh�es [EMAIL PROTECTED] wrote:
Hi guys I need some help with two things...
I have the following table:
CREATE TABLE `telephones` (
`contact_id` int(20) unsigned NOT NULL default '0',
`telephone_id` int(20) unsigned NOT NULL
Hi guys I need some help with two things...
I have the following table:
CREATE TABLE `telephones` (
`contact_id` int(20) unsigned NOT NULL default '0',
`telephone_id` int(20) unsigned NOT NULL default '0',
`telephone_country_code` char(5) NOT NULL default '',
`telephone_area_code`
I just revived a database that was in a version 3.23 server and moved it
to a 4.1 There are big fields of TEXT based data. They have a way of
compressing the amount of TEXT data by identifying common subchunks and
putting them in a subchunk table and replacing them with a marker
inside the
Thanks for your answer. It would certainly work provided having
enough disk space to do that. I thought something like
that but was hoping I can leverage fulltext and just
record the fulltext result between a each record
and each other record. Then I can group all records that
highly correlate
You can modify the algorithm I proposed to find groups of records that
are likely to have duplicate chunks. Simply record only a part of
hashes, something like: if md5(concat(word1,word2,...,word20))%32=0.
Disk usage for this table will be maybe 60 bytes per record, if your
average word is 8 bytes
Saqib Ali wrote:
Hello All,
What are best practices for deleting records in a DB. We need the
ability to restore the records.
Two obvious choices are:
1) Flag them deleted or undeleted
2) Move the deleted records to seperate table for deleted records.
We have a complex schema. However
Saqib Ali wrote:
Hello All,
What are best practices for deleting records in a DB. We need the
ability to restore the records.
Two obvious choices are:
1) Flag them deleted or undeleted
2) Move the deleted records to seperate table for deleted records.
We have a complex schema
Bastian Balthazar Bux wrote:
We need to track the modification to the records too so the route has
been to keep them all in a different, specular databases.
If the real table look like this:
CREATE TABLE `users` (
`id` int(11) NOT NULL auto_increment,
`ts` timestamp NOT NULL
default
Hello All,
What are best practices for deleting records in a DB. We need the
ability to restore the records.
Two obvious choices are:
1) Flag them deleted or undeleted
2) Move the deleted records to seperate table for deleted records.
We have a complex schema. However the the records
Saqib Ali wrote:
Hello All,
What are best practices for deleting records in a DB. We need the
ability to restore the records.
Two obvious choices are:
1) Flag them deleted or undeleted
2) Move the deleted records to seperate table for deleted records.
The first is what I like more.
While
Hello all,
I am using this script and it takes 100 % of the process, can anyone tell me
how to optimize this,
insert into incoming
select s.Date as Datein, s.Time as Timein, e.Date as Dateend, e.Time as
Timeend, s.CallingStationId, s.CalledStationId,
SEC_TO_TIME(unix_timestamp(concat(e.Date,'
on 11/09/2004 07:06:18 PM:
I am working on a data warehousing solution involving
mysql and have a question about best practices. We are
standardized on mysql 4.1, and this is for a rather
picky client.
We have a relational transaction database that stores
the results of customer calls
.
Eamon Daly
- Original Message -
From: Michael Haggerty [EMAIL PROTECTED]
To: [EMAIL PROTECTED]
Sent: Tuesday, November 09, 2004 6:06 PM
Subject: Best Practices
I am working on a data warehousing solution involving
mysql and have a question about best practices. We are
standardized on mysql
.
Eamon Daly
- Original Message -
From: Michael Haggerty [EMAIL PROTECTED]
To: [EMAIL PROTECTED]
Sent: Tuesday, November 09, 2004 6:06 PM
Subject: Best Practices
I am working on a data warehousing solution
involving
mysql and have a question about best
-
From: Michael Haggerty [EMAIL PROTECTED]
To: [EMAIL PROTECTED]
Sent: Wednesday, November 10, 2004 11:27 AM
Subject: Re: Best Practices
Yes, there can be a small lag in data updates, in fact
I believe the lag time will be less than a second
considering our architecture.
We have been
I am working on a data warehousing solution involving
mysql and have a question about best practices. We are
standardized on mysql 4.1, and this is for a rather
picky client.
We have a relational transaction database that stores
the results of customer calls and a dimensional
reporting database
server all together.
On Tue, 9 Nov 2004 16:06:18 -0800 (PST), Michael Haggerty
[EMAIL PROTECTED] wrote:
I am working on a data warehousing solution involving
mysql and have a question about best practices. We are
standardized on mysql 4.1, and this is for a rather
picky client.
We have
, and it currently takes
just over 3 minutes which is quite acceptable for what I'm doing. I'm
thinking about piping the output of mysqldump straight through gzip and then
ftp'd away to ease the disk access too, but that maybe later.
I would still like a best-practices guide though, so
In a message dated 2/11/2004 2:26:09 PM Eastern Standard Time,
[EMAIL PROTECTED] writes:
I read this over and over.. I am curious why replication is such high
finance?? I run it here. The Production system is a high finance machine and the
replicated box is a old clunker basically.. It doesn't
Hi,
I'd love to see this too. Even if it was a book that cost ?40 to buy, I'd
get a copy.
Hey, maybe someone can recommend a book - I've looked hard and not really
come up with anything better than the MySQL manual, which while great, is
missing the 'best practices' :(
Thanks,
Mike
Is there any best-practices wisdom on what is the most preferable
method of backing up moderately (~10-20,000 record) MySQL 4
databases? A mysql dump to store records as text, the
format provided
by the BACKUP sql command, or some other method? I am not asking
about replication
-Original Message-
From: Michael Collins [mailto:[EMAIL PROTECTED]
Is there any best-practices wisdom on what is the most preferable
method of backing up moderately (~10-20,000 record) MySQL 4
databases? A mysql dump to store records as text, the
format provided
From: David Brodbeck [EMAIL PROTECTED]
Sent: Wednesday, February 11, 2004 9:27 PM
-Original Message-
From: Michael Collins [mailto:[EMAIL PROTECTED]
Is there any best-practices wisdom on what is the most preferable
method of backing up moderately (~10-20,000 record) MySQL 4
-Original Message-
From: Madscientist [mailto:[EMAIL PROTECTED]
We use this mechanism, but we do our mysqldumps from a slave
so the time doesn't matter.
Excellent idea.
Interesting side effect: A GZIP of the data files is _huge_.
A GZIP of the
mysqldump is _tiny_. For our
. I'm
thinking about piping the output of mysqldump straight through gzip and then
ftp'd away to ease the disk access too, but that maybe later.
I would still like a best-practices guide though, so that if everything does
go wrong I'm sure that I've got everything I need to reconstruct the system
thing simply uses the MS scheduler in windows.
Might be a help
Paul
-Original Message-
From: Michael McTernan [mailto:[EMAIL PROTECTED]
Sent: 11 February 2004 21:41
To: David Brodbeck; Michael Collins
Cc: [EMAIL PROTECTED]
Subject: RE: best-practices backups
Hi,
I do just
PM, Michael Collins wrote:
Is there any best-practices wisdom on what is the most preferable
method of backing up moderately (~10-20,000 record) MySQL 4 databases?
A mysql dump to store records as text, the format provided by the
BACKUP sql command, or some other method? I am not asking about
Is there any best-practices wisdom on what is the most preferable
method of backing up moderately (~10-20,000 record) MySQL 4
databases? A mysql dump to store records as text, the format provided
by the BACKUP sql command, or some other method? I am not asking
about replication, rotating
We are implementing three or four MySql servers (as a start) and I'm
writing the Troubleshooting Guide for our operational staff. None of
these folks have any MySQL experience (and I'm a newbie myself). I need
a pretty basic 'Cheat Sheet' for troubleshooting common production type
problems.
The
We are implementing three or four MySql servers (as a start)
and I'm writing the Troubleshooting Guide for our operational
staff. None of these folks have any MySQL experience (and
I'm a newbie myself). I need a pretty basic 'Cheat Sheet' for
troubleshooting common production type
-Original Message-
From: Dan Muey [mailto:[EMAIL PROTECTED]
Sent: Wednesday, February 04, 2004 3:33 PM
To: Schwartz, Evelyn; [EMAIL PROTECTED]
Subject: RE: Is there any documentation of Best
Practices/Troubleshooting Guides for Administering MySQL
We are implementing three or four
On 4 Feb 2004, at 20:32, Dan Muey wrote:
We are implementing three or four MySql servers (as a start)
and I'm writing the Troubleshooting Guide for our operational
staff. None of these folks have any MySQL experience (and
I'm a newbie myself). I need a pretty basic 'Cheat Sheet' for
Dear All,
Please comment on Best Practices for sharing members database between
different portals.
Suppose we have 3 portals running on different networks.
Assignment is to make a single Login/Pass for all portals, means once
LogedIn in one of the portal, could able to access the other portals
On Tue, 09 Dec 2003 15:26:10 -0600 Tariq Murtaza [EMAIL PROTECTED]
wrote:
Please comment on Best Practices for sharing members database between
different portals.
Suppose we have 3 portals running on different networks.
Assignment is to make a single Login/Pass for all portals, means once
will
then set up a new 4016/data directory, and copy each database in for
testing. My question is : can someone give me some guidelines on the best
practices for setting this up? How to make the two installations completely
independant of each other? Also anything I should know about 4.016 thats
]
To: [EMAIL PROTECTED]
Sent: Thursday, November 06, 2003 8:35 AM
Subject: best practices for running 3.23 and 4.016 on the same box?
Hi all
I am planning an upgrade for a very busy 3.23 server to 4.016. As we dont
have a text box, I have to install both releases on the same machine, and
test
Hi list:
I would like to know some of the best practices to manage innodb tables. I
have an ibdata file that its size is 4.5GB, and it will increase every day
the max size of the hard disk is about 330GB, the question is should I
split this ibdata file in several files in a way that I can
Yes, if you have transaction logging turned on.
You can edit the transaction log, and run it against the restored database.
Subhakar Burri wrote:
Can I roll forward if I do backups using Mysqldump? Say, I did backups using Mysqldump @ 8:00 AM and my instance crashed @ 2:00 PM. I can restore the
On Fri, Jun 27, 2003 at 08:08:40AM -0500, gerald_clark wrote:
Yes, if you have transaction logging turned on.
You can edit the transaction log, and run it against the restored database.
MyISAM doesn't have transactions.
Jeremy
--
Jeremy D. Zawodny | Perl, Web, MySQL, Linux Magazine,
Ok, update log.
Jeremy Zawodny wrote:
On Fri, Jun 27, 2003 at 08:08:40AM -0500, gerald_clark wrote:
Yes, if you have transaction logging turned on.
You can edit the transaction log, and run it against the restored database.
MyISAM doesn't have transactions.
Jeremy
--
MySQL General
question of can I roll forward the changes to both table types ... Am I
missing something? Please clarify
SB
Original Message-
From: gerald_clark [mailto:[EMAIL PROTECTED]
Sent: Friday, June 27, 2003 11:39 AM
To: [EMAIL PROTECTED]
Subject: Re: Best Practices for mySQL Backups in Enterprise
Ok
Hi Gerald,
Do you know some good information about it, seems like I need to brush up a
bit on this.
I dont understand how you want to do a roll forward for a MySQL table -
especially if the backup is lets ay from 8 AM and the crash is at 2 PM.
Best regards
NIls Valentin
Tokyo/Japan
2003 6
. Ivey; [EMAIL PROTECTED]
Subject: Re: Best Practices for mySQL Backups in Enterprise
Hi Gerald,
Do you know some good information about it, seems like I need to brush up a
bit on this.
I dont understand how you want to do a roll forward for a MySQL table -
especially if the backup is lets ay
Hi Subhakar,
I would be interested to know what you mean with roll forward ?
In case you have another backup let's say @10AM you could use this one, but if
you dont have another backup where do you want to do a roll forward from ??
Do I miss something here ??
Best regards
Nils Valentin
Imro STROK [EMAIL PROTECTED] writes:
But I would also like to have some MySQL Best Practices documents
regarding:
* Performance Tuning
* Backup Recovery
Appreciate if you guys/gals can send me some documents.
Take a look at URL:
http://www.onlamp.com/pub/a/onlamp/2002/07/11
the book Teach Yourself MySQL in 21 days
by Mark Maslakowski
is very helpful to speed our MySQL skills.
But I would also like to have some MySQL Best Practices documents
regarding:
* Performance Tuning
* Backup Recovery
Appreciate if you guys/gals can send me some documents.
Best Regards
Hello,
I'm trying to determine the best way to setup a new mysql install. I have
about 100 clients and each client has the same 16 tables. In a years time
there will be about 3.6 million total rows spread out between the tables.
The tables will be updated nightly from a legacy system. All other
* Stephen S Zappardo
a) 1 db with 16 tables b) 100 dbs each with 16 tables
* adam nelson
Certainly 1 db with 16 tables.
Why? Normally, bigger means slower...
--
Roger
-
Before posting, please check:
* adam nelson
Management seems like the biggest reason for me. Just from a time spent
point of view, I would go with 16 tables instead of 1600. Not only
that, I wonder if there would be a big memory hit from having all those
objects open at once. Just seems to me that mysql was designed
54 matches
Mail list logo