Were you using -mx and -ms (setting heap size ?)

  Cheers,
   Winton

>hi to all,
>
>please help! I think I mixed my brain up already with this stuff...
>
>I'm trying to index about 29 textfiles where the biggest one is ~700Mb and
>the smallest ~300Mb. I achieved once to run the whole index, with a merge
>factor = 10 and maxMergeDocs=10000. This took more than 35 hours I think
>(don't know exactly) and it didn't use much RAM (though it could have).
>unfortunately I had a call to optimize at the end and while optimization an
>IOException (File to big) occured (while merging).
>
>As I run the program on a multi-processor machine I now changed the code to
>index each file in a single thread and write to one single IndexWriter. the
>merge factor is still at 10. maxMergeDocs is at 1.000.000. I set the maximum
>heap size to 1MB.
>
>I tried to use RAMDirectory (as mentioned in the mailing list) and just use
>IndexWriter.addDocument(). At the moment it seems not to make any difference.
>after a while _all_ the threads exit one after another (not all at once!)
>with an OutOfMemoryError. the priority of all of them is at the minimum.
>
>even if the multithreading doesn't increase performance I would be glad if I
>could just once get it running again.
>
>I would be even happier if someone could give me a hint what would be the
>best way to index this amount of data. (the average size of an entry that
>gets parsed for a Document is about 1Kb.)
>
>thanx for any help!
>chantal
>
>--
>To unsubscribe, e-mail:   <mailto:[EMAIL PROTECTED]>
>For additional commands, e-mail: <mailto:[EMAIL PROTECTED]>


Winton Davies
Lead Engineer, Overture (NSDQ: OVER)
1820 Gateway Drive, Suite 360
San Mateo, CA 94404
work: (650) 403-2259
cell: (650) 867-1598
http://www.overture.com/

--
To unsubscribe, e-mail:   <mailto:[EMAIL PROTECTED]>
For additional commands, e-mail: <mailto:[EMAIL PROTECTED]>

Reply via email to