Hi guys Thank you for your answers.
Just to clarify the reasons for my question: - I am at the stage where I write my code, so I have to run my unit tests like 50 times a day :) Hence, every second shaven off from the start time is invaluable. Yes, I agree with Jörn, the lifecycle of the model is indeed the same with the one of my application. - I am already ussing a decent SSD, so loading time as a factor is off-the-table. I will keep tinker around it and let you know how it goes ... Many thanks. Ioan On Sat, Sep 7, 2013 at 9:21 PM, Aliaksandr Autayeu <[email protected]>wrote: > Ioan, often hard drive speed limits you more than processor speed. That is, > it might be faster to load 5M from disk and unpack than load unpacked > 25M from disk. > > Aliaksandr > > > On Wed, Sep 4, 2013 at 11:59 AM, Jörn Kottmann <[email protected]> wrote: > > > On 08/26/2013 03:00 PM, Ioan Barbulescu wrote: > > > >> Hi guys > >> > >> Short question, please: > >> > >> Currently, the opennlp models come as zipped files. > >> > >> Is it possible to use them in an expanded / un-zipped form? > >> (and how?) > >> > >> Zipped is very neat and clean, but it adds some time when reading the > >> file. > >> I am interested in speeding up as much as possible the load time. > >> > > > > You can probably repackage the zip files without using compression. > > Anyway I doubt that it adds much time, did you profile the loading code? > > > > As far as I know is the slowest part to build the maxent model, maybe > that > > can be speed up, I never profiled that part of OpenNLP. > > > > The life-cycle of a model should be the same as of your application, > > maybe you can > > just find a way to reuse them, instead of loading them over and over > again. > > > > Jörn > > > > > > >
