I have a business issue where I need to store and eventually update 30000 complex xml documents (about 60 k each, and validated against the same schema). We are extracting data from a legacy system, and we would like to create xml documents, so that we can eventually tranmit the data to another system. Is it feasible to use an XML database for a dataset of this size? In the tests I have done so far, indexing seems to be a major issue. For 30000 documents, it takes several hours to create an index on one element, and the index is very large. I created the index after adding all 30000 (test) documents.

Here are some general questions:

For a dataset of this size, will updating be an issue?

Should we store all documents as part of one collection?

Should we break up the documents into component parts, and reassemble them when we need to? (for query performance reasons)

Are there any indexing strategies that offer optimal performance?

Thanks very much,
Bob Corcoran










_________________________________________________________________
MSN 8 with e-mail virus protection service: 2 months FREE* http://join.msn.com/?page=features/virus




Reply via email to