Hello there !
It seems to me that there is a storage problem, when you create a lot of nodes,
one by one, using this algorithm :
1. for each node to create
a. create node
b. fill node properties/child nodes
c. save session
2. end for
The default_node and default_prop tables number of rows (and size) increases
very fast, and in an unacceptable way.
I had a 35 million default_node table after inserting like this 27 000 nodes in
a repository.
Then I used the other algorithm :
1. for each node to create
a. create node
b. fill node properties/child nodes
2. end for
3. save session
And this gives a much better situation (currently I have a 36 000 content
repository, and my tables are correct - 60 000 rows for node table,
576 000 rows for properties).
The problem here is that in a production environment, users are going to create
their nodes one by one, day after day, never by full blocks.
So is there a storage problem ?
Frederic Esnault