Are you are indexing multiple documents? If so, split them into multiple files. A single XML file with all documents is not a good idea. Solr is designed to
use batches for indexing.

It will be extremely hard to index a 1TB XML file. I would guess that would need
a JVM heap of well over 1TB.

wunder

On Oct 8, 2009, at 6:56 AM, Noble Paul നോബിള്‍ नोब्ळ् wrote:

you can write a simple program which streams the file from the disk to
post it to Solr


On Thu, Oct 8, 2009 at 7:10 PM, Elaine Li <elaine.bing...@gmail.com> wrote:
You can increase the java heap size, e.g. java -Xms128m -Xmx8192m - jar <*.xml>
Or i split the file if it is too big.

Elaine

On Thu, Oct 8, 2009 at 6:47 AM, Pravin Karne
<pravin_ka...@persistent.co.in> wrote:
Hi,
I am new to solr. I am able to index, search and update with small size(around 500mb) But if I try to index file with 5 to 10 or more that (500mb) it gives memory heap exception. While investigation I found that post jar or post.sh load whole file in memory.

I use one work around with dividing small file in small files..and it's working

Is there any other way to post large file as above work around is not feasible for 1 TB file

Thanks
-Pravin


DISCLAIMER
==========
This e-mail may contain privileged and confidential information which is the property of Persistent Systems Ltd. It is intended only for the use of the individual or entity to which it is addressed. If you are not the intended recipient, you are not authorized to read, retain, copy, print, distribute or use this message. If you have received this communication in error, please notify the sender and delete all copies of this message. Persistent Systems Ltd. does not accept any liability for virus infected mails.





--
-----------------------------------------------------
Noble Paul | Principal Engineer| AOL | http://aol.com


Reply via email to