You know, sometimes simple and fast is a good way to do things. Do you have an auto-increment PK in that table? Would be helpful. As for Cayenne, can you can flush (invalidate) any active DataContexts (at least the objects for that table) when the load occurs?
/dev/mrg On 12/5/06, Peter Schröder <[EMAIL PROTECTED]> wrote:
we are deleting all rows with truncate table first. then loading cvs with load data infile. i would prefer not to use this method, but it is simple and fast. -----Ursprüngliche Nachricht----- Von: Michael Gentry [mailto:[EMAIL PROTECTED] Gesendet: Dienstag, 5. Dezember 2006 14:38 An: [email protected] Betreff: Re: updating large number of data Are you deleting all of the original data and then doing inserts or are you doing updates? Thanks, /dev/mrg On 12/5/06, Peter Schröder <[EMAIL PROTECTED]> wrote: > hi, > > we get a cvs-file with a large number of user-data every hour. and we want to replace the existing data in our database with that. is there a best-practice to do something like this? > currently we are doing that with php an using mysql load data infile with the cvs-file. but i think that doing this with cayenne would leave the context in a bad state. > > hava a nice day, > peter >
