Actually this will be a part of the application logic. We have real time weather data which we constantly receive and insert into the DB. - Derek
________________________________ From: Mamta Satoor [mailto:[EMAIL PROTECTED] Sent: Wednesday, March 14, 2007 5:02 PM To: Derby Discussion Subject: Re: Large multi-record insert performance If this bulk insert is not normal part of application logic and only done once in a while then I wonder if import using SYSCS_UTIL.SYSCS_IMPORT _TABLE would be a faster way to load data. Mamta On 3/14/07, Lance J. Andersen <[EMAIL PROTECTED]> wrote: Mike Matrigali wrote: > > > Lance J. Andersen wrote: >> >> >> Even if the backend does not provide optimization for batch >> processing, i would hope that there would be still some efficiency >> especially in a networked environment vs building the strings, >> invoking execute() 1000 times in the amount of data on the wire... >> >> > I could not tell from the question whether this was network or not. I > agree in network then limiting execution probably is best. In embedded > I am not sure - I would not be surprised if doing 1000 in batch is > slower than just doing the executes. > > In either case I really would stay away from string manipulation as much > as possible and also stay away from things that create very long SQL > statements like 1000 term values clauses. i agree completely. Let the driver do the heavy lifting :-)
