I have a business issue where I need to store and eventually update 30000
complex xml documents (about 60 k each, and validated against the same
schema). We are extracting data from a legacy system, and we would like to
create xml documents, so that we can eventually tranmit the data to another
system. Is it feasible to use an XML database for a dataset of this size? In
the tests I have done so far, indexing seems to be a major issue. For 30000
documents, it takes several hours to create an index on one element, and the
index is very large. I created the index after adding all 30000 (test)
documents.
Here are some general questions:
For a dataset of this size, will updating be an issue?
Should we store all documents as part of one collection?
Should we break up the documents into component parts, and reassemble them
when we need to? (for query performance reasons)
Are there any indexing strategies that offer optimal performance?
Thanks very much,
Bob Corcoran
_________________________________________________________________
MSN 8 with e-mail virus protection service: 2 months FREE*
http://join.msn.com/?page=features/virus