Sedillo, Derek (Mission Systems) wrote:
Hello,
I have several large Oracle tables (2+ million records) which I need to import data from to Derby. Here is what I have tried so far: 1. I have dumped the data to comma separated flat files.
2.  Used the import table utility like this:
/CALL SYSCS_UTIL.SYSCS_IMPORT_TABLE (null,'SSJ_CNT','SSJ_CNT.csv',null, null,null,0);/


Import will run slow, if the table has triggers, foreing key references or if the table already has some data. These force the data to be logged , which makes import run slow.

If you have foreign keys/triggers you may want drop them and add them back after the data is imported. If you do that import might run faster.



3. After '4 hours' of running it appears to have frozen up. There is both a db.lck and dbex.lck file present which I will delete soon.

These files are not related to import. You should not be deleting them. They are used to prevent multiple jvm instances booting
the same database concurrenly.


Do you think that 2 million records is just too much for this utility to handle? Is there a better way to transfer data from another database besides the IMPORT_TABLE utility?

I have not tried with 2 million rows , IMPORT_TABLE is the fastest way to tranfer data from files into a derby database.



hope that helps
-suresh

Reply via email to