On 12-02-16 18:26, Erik Sundvall wrote:
if you are experimenting with open source native XML DBs for openEHR,
it preformed well for "clinical" patient-specific querying even though
all xml databases we tested were not suitable for ad hoc
epidemiological population queries (without query specific indexing)
A very interesting paper. I have some first opinions on that. But first
I need to explain what I think about the matter.
I have not prepared the story below, so there may be things which I
write to fast. See it as provisional view, not as a hard opinion.
---------
There are relational database-configurations for OLAP and for OLTP. The
combination is hard to find. There are reasons. This is a classic problem.
You need specific indexes for data-mining (OLAP), and for every extra
data-mining query you need more indexes, especially if you don't have
time to wait a night for the result. Those extra indexes stand in the
way for transactional processing (OLTP) because they need to be updated,
and that is unnecessary burden for the OLTP-processes, as longer as the
database exist, the burden becomes heavier.
That is why OLAP and OLTP are not often combined in one configuration.
So many professional databases have extra features for OLAP, I worked,
years ago with Oracle for a big Dutch television company, and my main
job was to create indexes for marketing purposes.
We ran those unusual queries during the night and stored the result in
special tables, Oracle called them "materialized views".
The day after, those views were processed in analyzing software, like
SPSS, and after that, thrown away.
It was a database with 900,000 persons in it, and every person had a lot
of history of web-history, personal interests, etc.
"How much interest does a person have for opera, and is he able to pay
for opera, is it worth to call him for a ticket-offer, we cannot call
900,000 persons"
These were complex queries based on things the customer bought,
television programs he was interested in, web-activities.
That was the kind of thing they did with the database.
So this could well be compared with a medical database, regarding to
size and complexity.
The same difficulties count for XML databases. That is why XML databases
have also features for creating extra indexes.
Oracle, by the way, if it knows the structure of XML (via XSD), it
breaks, underwater, XML into relational data, and store it in a
relational database. It also converts XQuery to SQL.
In this way, it has the weakness and advantages of a relational
database, and it needs the extra indexes for unusual queries, but on the
developer view it is XML.
-------
Comparing XML and relational, for OpenEHR, I favor XML, because it can
easily reflect the structures which need to be stored. It makes the
data-processing layer less complex. There is a lot of tooling around
XML, XML-schema to make the database-structure known to Oracle,
Schematron to validate against archetypes. This is very matured
software, and therefor the complexity is solved years ago, and well
tested. It is hidden complexity, and matured hidden complexity is no
problem.
--------
And if you want to do data-mining, like epidemiological research, and
you have the time to plan the research, then the classical database, XML
or RDB, is OK.
In my opinion, there is not often a real need for adhoc data-mining
(epidemiological research) queries, with result in a few minutes. They
are always planned, and creating the indexes and storing the result in
"materialized views" are part of the work one has to do for data-mining
research on data.
So, I don't think there is a real need for this.
--------
Regarding to XML databases, Oracle has a solution, which can perform
well if it is professionally maintained.
This is often a point, because professional Oracle maintaining regarding
to advanced use is very expensive.
Another company is MarkLogic. It is said that MarkLogic is better, but I
don't know that from own experience.
Both are free to use for developers.
You must think of numbers to 35,000 Euro a year for licenses, which is
not very much for a big hospital, but very much for a small health service
The open source XML ExistDB database is not very good for data-mining,
is my personal experience.
So, we must ask ourselves, are we solving a problem that no one experiences?
--------
There are a few advantages to OpenEHR. Data are immutable, never
changed, never deleted. This makes a few difficult steps unnecessary.
The Dewey concepts look very attractive, although it is also created
with deleting and changing data in mind.
This is very important for normal company use.
But, as said, we don't have that in OpenEHR. Medical data always grow,
it are always new data. An event that has passed will never change.
The only things that change (in OpenEHR they are versioned, but from the
user perspective, they change), that are demographic data. And one can
live with that, create extra provisions for that demographic
database-section, which is only a small part of the complete database.
Often, the demographic database is external anyway.
So, my thoughts, maybe Dewey is too good.
Path-values (to leaf-nodes) storage is enough
The paths, combined with ID's are keys, and are much alike XPath, so it
is easy to store XML in a path-value database.
And querying is also easy, because all queries are path based.
I think, for OpenEHR this is the fastest solution. But maybe I overlook
something.
Maybe I say something stupid. I am not offended if you say so (maybe in
other words ;-).
We all want and need to learn.
What we still need to do to build this solution is handle the XQuery
grammar and let it run on path-based-database.
This is not very easy, but also not very hard. Maybe the algorithms are
already to find.
Like the Dewey algorithm this can run on any database, also on free open
source databases.
I think you get an excellent performance on Postgres.
A path-value database is easy to index, it only needs a few. The
inserting will stay very fast, always.
Lets do some calculations, for fun:
How many path-value-combinations do you have to store for an average
composition?
Maybe 30? How many compositions are for an average patient? 10,000?
So every patient needs 300,000 path-values. So you can store 10,000
patients in 3 billion records.
This is not much for Postgres, and the simple indexes needed.
When you need to store 900,000 patients you need 90 separate tables.
Very cheap, very fast, also for adhoc queries, and easy to accomplish, I
think.
I am very interested in opinions on this.
Thanks
Bert
_______________________________________________
openEHR-technical mailing list
openEHR-technical@lists.openehr.org
http://lists.openehr.org/mailman/listinfo/openehr-technical_lists.openehr.org