Have you found an explanation to that?

On Tue, Feb 17, 2015 at 12:12 PM, Markus Jelsma <markus.jel...@openindex.io>
wrote:

> We have seen an increase between 4.8.1 and 4.10.
>
> -----Original message-----
> > From:Dmitry Kan <solrexp...@gmail.com>
> > Sent: Tuesday 17th February 2015 11:06
> > To: solr-user@lucene.apache.org
> > Subject: unusually high 4.10.2 vs 4.3.1 RAM consumption
> >
> > Hi,
> >
> > We are currently comparing the RAM consumption of two parallel Solr
> > clusters with different solr versions: 4.10.2 and 4.3.1.
> >
> > For comparable index sizes of a shard (20G and 26G), we observed 9G vs
> 5.6G
> > RAM footprint (reserved RAM as seen by top), 4.3.1 being the winner.
> >
> > We have not changed the solrconfig.xml to upgrade to 4.10.2 and have
> > reindexed data from scratch. The commits are all controlled on the
> client,
> > i.e. not auto-commits.
> >
> > Solr: 4.10.2 (high load, mass indexing)
> > Java: 1.7.0_76 (Oracle)
> > -Xmx25600m
> >
> >
> > Solr: 4.3.1 (normal load, no mass indexing)
> > Java: 1.7.0_11 (Oracle)
> > -Xmx25600m
> >
> > The RAM consumption remained the same after the load has stopped on the
> > 4.10.2 cluster. Manually collecting the memory on a 4.10.2 shard via
> > jvisualvm dropped the used RAM from 8,5G to 0,5G. But the reserved RAM as
> > seen by top remained at 9G level.
> >
> > This unusual spike happened during mass data indexing.
> >
> > What else could be the artifact of such a difference -- Solr or JVM? Can
> it
> > only be explained by the mass indexing? What is worrisome is that the
> > 4.10.2 shard reserves 8x times it uses.
> >
> > What can be done about this?
> >
> > --
> > Dmitry Kan
> > Luke Toolbox: http://github.com/DmitryKey/luke
> > Blog: http://dmitrykan.blogspot.com
> > Twitter: http://twitter.com/dmitrykan
> > SemanticAnalyzer: www.semanticanalyzer.info
> >
>



-- 
Dmitry Kan
Luke Toolbox: http://github.com/DmitryKey/luke
Blog: http://dmitrykan.blogspot.com
Twitter: http://twitter.com/dmitrykan
SemanticAnalyzer: www.semanticanalyzer.info

Reply via email to