thanks, I saw that.  I wondered if there was some obvious thing I could
check before starting rewriting the xslt (I know, I'm lazy)


Eric


-----Message d'origine-----
De : Ralph Goers [mailto:[EMAIL PROTECTED]
Envoyé : 28 octobre, 2005 11:51
À : users@cocoon.apache.org
Objet : Re: processing large files


Boisvert, Éric wrote:

>Hi all
>
>I need to process large xml file and as I tested with increasingly larger
>file, the time to process suddently increased a lot.  For instance, 200 K
>files took 0.8 seconds, 400 K file 2.5 sec and when I get near 1 Meg, it
>jumps to 30 seconds (nearly 10 times, for twice the size).. I played with
>the pipeline caching, outputBufferSize, etc.. even boosted CATALINA_OPTS to
>512 Megs, nothing helped.  I guess this is related to the fact that at some
>point the incoming document cannot be loaded entirely in memory.
>
>Anyone has an idea to fix this ?
>
>Cheers and thanks
>
>  
>
This was the subject of one of the presentations at the Cocoon 
GetTogether (http://www.cocoongt.org).  Here is a link to the 
presentation. 
http://cocoongt.hippo12.castaserver.com/cocoongt/nico-verwer-performance.pdf

Ralph


---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]