Specifically answering the _indexing_ part of the question, in solrconfig.xml theres a ramBufferSizeMB (from memory) that governs how much RAM is used while indexing before flushing to disk. I think the default is 100MB or so.
On Thu, May 11, 2017 at 2:34 PM, Rick Leir <rl...@leirtech.com> wrote: > You could limit the Java heap, but that is counter productive. You should > have a look at how much heap it uses. But let Solr use what it needs. My > guess is that your -Xmx or -Xmm is too low at the moment. > > Apart from that, Solr will mmap large files. When there is not enough RAM for > this, any swapping will be bad for performance. When you see this it will be > time to shard your Solr core or buy more Ram. > Cheers -- Rick > > > > On May 11, 2017 2:26:23 PM EDT, Deepak Mali <deepak3...@gmail.com> wrote: >>Hello All, >> >>if there is any way to set threshold memory to the solr indexing >>process. >>My computer is hung and the indexing process is killed by the OS. >> >>So , I was wondering if there is any way to set threshold memory usage >>to >>solr indexing process in linux environments. >> >>Thank you in advance. > > -- > Sorry for being brief. Alternate email is rickleir at yahoo dot com