On Jun 10, 2005, at 9:33 AM, Chris Collins wrote:

How many documents did you try to index?

Only about 4000 at the moment.

  I am using a relatively large
minMergeDoc that causes me to run out of memory when I make such a change. (I
am using 1/2 gb of heap btw).

I was running out of memory as well until I gave Java a larger heap to work with. I am assuming that a dedicated indexing machine (as well as search) is going to need a mountain of memory. I figure I will be giving Java gigs to play with.

I believe changing it in the outputstream object
means that a lot of in memory only objects use that size too.

This I need to look into. At a guess, I would think that there would be an OutputStream object for each open segment, and each file in that segment. A consolidated index *might* use less but of course we are trying to improve performance here, and the consolidated index does incur a cost. Assuming 10 segments and 10 files within each segment, that's 100 OutputStream objects or 809,600 bytes. That'll grow quickly with merge tweaks. Those larger writes do save a bunch of system calls and make (maybe) better use of your filers block size. This grows quickly with maxMerge tweaks. Of course this could be utterly incorrect, I need to look into this a bit more carefully.

I dont know I would of used truss in this regard, this only points out what size hit the kernel not what went over the wire. I would suggest using
ethereal to ensure thats how its ending up on the wire.

True, hadn't gotten that far yet. :-)

Peter



---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to