Hello,
Since you use caching, you can monitor the eviction parameter on the solr
admin page (http://localhost:port/solr/admin/stats.jsp#cache). If it is non
zero, the cache can be made e.g. bigger.
queryResultWindowSize=50 in my case.
Not sure, if solr 3.1 supports, but in 1.4 I have:
<HashDocSet maxSize="1000" loadFactor="0.75"/>

Does the OOM happen on update/commit or search?

Dmitry

On Wed, Sep 14, 2011 at 2:47 PM, Rohit <ro...@in-rev.com> wrote:

> Thanks Dmirty for the offer to help, I am using some caching in one of the
> cores not. Earlier I was using on other cores too, but now I have commented
> them out because of frequent OOM, also some warming up in one of the core. I
> have share the links for my config files for all the 4 cores,
>
> http://haklus.com/crssConfig.xml
> http://haklus.com/rssConfig.xml
> http://haklus.com/twitterConfig.xml
> http://haklus.com/facebookConfig.xml
>
>
> Thanks again
> Rohit
>
>
> -----Original Message-----
> From: Dmitry Kan [mailto:dmitry....@gmail.com]
> Sent: 14 September 2011 10:23
> To: solr-user@lucene.apache.org
> Subject: Re: Out of memory
>
> Hi,
>
> OK 64GB fits into one shard quite nicely in our setup. But I have never
> used
> multicore setup. In total you have 79,9 GB. We try to have 70-100GB per
> shard with caching on. Do you do warming up of your index on starting?
> Also,
> there was a setting of pre-populating the cache.
>
> It could also help, if you can show some parts of your solrconfig file.
> What
> is the solr version you use?
>
> Regards,
> Dmitry
>
> On Wed, Sep 14, 2011 at 11:38 AM, Rohit <ro...@in-rev.com> wrote:
>
> > Hi Dimtry,
> >
> > To answer your questions,
> >
> > -Do you use caching?
> > I do user caching, but will disable it and give it a go.
> >
> > -How big is your index in size on the disk?
> > These are the size of the data folder for each of the cores.
> > Core1 : 64GB
> > Core2 : 6.1GB
> > Core3 : 7.9GB
> > Core4 : 1.9GB
> >
> > Will try attaching a jconsole to my solr as suggested to get a better
> > picture.
> >
> > Regards,
> > Rohit
> >
> >
> > -----Original Message-----
> > From: Dmitry Kan [mailto:dmitry....@gmail.com]
> > Sent: 14 September 2011 08:15
> > To: solr-user@lucene.apache.org
> > Subject: Re: Out of memory
> >
> > Hi Rohit,
> >
> > Do you use caching?
> > How big is your index in size on the disk?
> > What is the stack trace contents?
> >
> > The OOM problems that we have seen so far were related to the
> > index physical size and usage of caching. I don't think we have ever
> found
> > the exact cause of these problems, but sharding has helped to keep each
> > index relatively small and OOM have gone away.
> >
> > You can also attach jconsole onto your SOLR via the jmx and monitor the
> > memory / cpu usage in a graphical interface. I have also run garbage
> > collector manually through jconsole sometimes and it was of a help.
> >
> > Regards,
> > Dmitry
> >
> > On Wed, Sep 14, 2011 at 9:10 AM, Rohit <ro...@in-rev.com> wrote:
> >
> > > Thanks Jaeger.
> > >
> > > Actually I am storing twitter streaming data into the core, so the rate
> > of
> > > index is about 12tweets(docs)/second. The same solr contains 3 other
> > cores
> > > but these cores are not very heavy. Now the twitter core has become
> very
> > > large (77516851) and its taking a long time to query (Mostly facet
> > queries
> > > based on date, string fields).
> > >
> > > After sometime about 18-20hr solr goes out of memory, the thread dump
> > > doesn't show anything. How can I improve this besides adding more ram
> > into
> > > the system.
> > >
> > >
> > >
> > > Regards,
> > > Rohit
> > > Mobile: +91-9901768202
> > > About Me: http://about.me/rohitg
> > >
> > > -----Original Message-----
> > > From: Jaeger, Jay - DOT [mailto:jay.jae...@dot.wi.gov]
> > > Sent: 13 September 2011 21:06
> > > To: solr-user@lucene.apache.org
> > > Subject: RE: Out of memory
> > >
> > > numDocs is not the number of documents in memory.  It is the number of
> > > documents currently in the index (which is kept on disk).  Same goes
> for
> > > maxDocs, except that it is a count of all of the documents that have
> ever
> > > been in the index since it was created or optimized (including deleted
> > > documents).
> > >
> > > Your subject indicates that something is giving you some kind of Out of
> > > memory error.  We might better be able to help you if you provide more
> > > information about your exact problem.
> > >
> > > JRJ
> > >
> > >
> > > -----Original Message-----
> > > From: Rohit [mailto:ro...@in-rev.com]
> > > Sent: Tuesday, September 13, 2011 2:29 PM
> > > To: solr-user@lucene.apache.org
> > > Subject: Out of memory
> > >
> > > I have solr running on a machine with 18Gb Ram , with 4 cores. One of
> the
> > > core is very big containing 77516851 docs, the stats for searcher given
> > > below
> > >
> > >
> > >
> > > searcherName : Searcher@5a578998 main
> > > caching : true
> > > numDocs : 77516851
> > > maxDoc : 77518729
> > > lockFactory=org.apache.lucene.store.NativeFSLockFactory@5a9c5842
> > > indexVersion : 1308817281798
> > > openedAt : Tue Sep 13 18:59:52 GMT 2011
> > > registeredAt : Tue Sep 13 19:00:55 GMT 2011
> > > warmupTime : 63139
> > >
> > >
> > >
> > > .         Is there a way to reduce the number of docs loaded into
> memory
> > > for
> > > this core?
> > >
> > > .         At any given time I dont need data more than past 15 days,
> > unless
> > > someone queries for it explicetly. How can this be achieved?
> > >
> > > .         Will it be better to go for Solr replication or distribution
> if
> > > there is little option left
> > >
> > >
> > >
> > >
> > >
> > > Regards,
> > >
> > > Rohit
> > >
> > > Mobile: +91-9901768202
> > >
> > > About Me:  <http://about.me/rohitg> http://about.me/rohitg
> > >
> > >
> > >
> > >
> >
> >
>
>
> --
> Regards,
>
> Dmitry Kan
>
>


-- 
Regards,

Dmitry Kan

Reply via email to