Hello: Any follow up? 2016-11-03 11:18 GMT-07:00 Ray Niu <newry1...@gmail.com>:
> the soft commit is 15 seconds and hard commit is 10 minutes. > > 2016-11-03 11:11 GMT-07:00 Erick Erickson <erickerick...@gmail.com>: > >> Followup question: You say you're indexing 100 docs/second. How often >> are you _committing_? Either >> soft commit >> or >> hardcommit with openSearcher=true >> >> ? >> >> Best, >> Erick >> >> On Thu, Nov 3, 2016 at 11:00 AM, Ray Niu <newry1...@gmail.com> wrote: >> > Thanks Joel >> > here is the information you requested. >> > Are you doing heavy writes at the time? >> > we are doing write very frequently, but not very heavy, we will update >> > about 100 solr document per second. >> > How many concurrent reads are are happening? >> > the concurrent reads are about 1000-2000 per minute per node >> > What version of Solr are you using? >> > we are using solr 5.5.2 >> > What is the field definition for the double, is it docValues? >> > the field definition is >> > <field name="fieldName" type="tdouble" indexed="true" stored="true" >> > docValues="true"/> >> > >> > >> > 2016-11-03 6:30 GMT-07:00 Joel Bernstein <joels...@gmail.com>: >> > >> >> Are you doing heavy writes at the time? >> >> >> >> How many concurrent reads are are happening? >> >> >> >> What version of Solr are you using? >> >> >> >> What is the field definition for the double, is it docValues? >> >> >> >> >> >> >> >> >> >> Joel Bernstein >> >> http://joelsolr.blogspot.com/ >> >> >> >> On Thu, Nov 3, 2016 at 12:56 AM, Ray Niu <newry1...@gmail.com> wrote: >> >> >> >> > Hello: >> >> > We are using export handler in Solr Cloud to get some data, we >> only >> >> > request for one field, which type is tdouble, it works well at the >> >> > beginning, but recently we saw high CPU issue in all the solr cloud >> >> nodes, >> >> > we took some thread dump and found following information: >> >> > >> >> > java.lang.Thread.State: RUNNABLE >> >> > >> >> > at java.lang.Thread.isAlive(Native Method) >> >> > >> >> > at >> >> > org.apache.lucene.util.CloseableThreadLocal.purge( >> >> > CloseableThreadLocal.java:115) >> >> > >> >> > - locked <0x00000006e24d86a8> (a java.util.WeakHashMap) >> >> > >> >> > at >> >> > org.apache.lucene.util.CloseableThreadLocal.maybePurge( >> >> > CloseableThreadLocal.java:105) >> >> > >> >> > at >> >> > org.apache.lucene.util.CloseableThreadLocal.get( >> >> > CloseableThreadLocal.java:88) >> >> > >> >> > at >> >> > org.apache.lucene.index.CodecReader.getNumericDocValues( >> >> > CodecReader.java:143) >> >> > >> >> > at >> >> > org.apache.lucene.index.FilterLeafReader.getNumericDocValues( >> >> > FilterLeafReader.java:430) >> >> > >> >> > at >> >> > org.apache.lucene.uninverting.UninvertingReader.getNumericDocValues( >> >> > UninvertingReader.java:239) >> >> > >> >> > at >> >> > org.apache.lucene.index.FilterLeafReader.getNumericDocValues( >> >> > FilterLeafReader.java:430) >> >> > >> >> > Is this a known issue for export handler? As we only fetch up to 5000 >> >> > documents, it should not be data volume issue. >> >> > >> >> > Can anyone help on that? Thanks a lot. >> >> > >> >> >> > >