Which java version? 32 or 64 bit? You start Solr with default 512Mb heap size? Other software running on the machine?
-- Jan Høydahl, search solution architect Cominvent AS - www.cominvent.com > 25. jan. 2019 kl. 13:05 skrev Zheng Lin Edwin Yeo <edwinye...@gmail.com>: > > Hi Jan and Shawn, > > For your info, this is another debug query. > > "debug":{ > > "rawquerystring":"johnny", > > "querystring":"johnny", > > "parsedquery":"searchFields_tcs:johnny", > > "parsedquery_toString":"searchFields_tcs:johnny", > > "explain":{ > > "192280":"\n12.8497505 = weight(searchFields_tcs:johnny in > 75730) [SchemaSimilarity], result of:\n 12.8497505 = > score(doc=75730,freq=4.0 = termFreq=4.0\n), product of:\n 7.5108404 > = idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq + > 0.5)) from:\n 473.0 = docFreq\n 865438.0 = docCount\n > 1.7108272 = tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 - > b + b * fieldLength / avgFieldLength)) from:\n 4.0 = > termFreq=4.0\n 1.2 = parameter k1\n 0.75 = parameter b\n > 26.66791 = avgFieldLength\n 25.0 = fieldLength\n”, > > "QParser":"LuceneQParser", > > "timing":{ > > "time":350.0, > > "prepare":{ > > "time":0.0, > > "query":{ > > "time":0.0}, > > "facet":{ > > "time":0.0}, > > "facet_module":{ > > "time":0.0}, > > "mlt":{ > > "time":0.0}, > > "highlight":{ > > "time":0.0}, > > "stats":{ > > "time":0.0}, > > "expand":{ > > "time":0.0}, > > "terms":{ > > "time":0.0}, > > "debug":{ > > "time":0.0}}, > > "process":{ > > "time":348.0, > > "query":{ > > "time":287.0}, > > "facet":{ > > "time":0.0}, > > "facet_module":{ > > "time":0.0}, > > "mlt":{ > > "time":0.0}, > > "highlight":{ > > "time":54.0}, > > "stats":{ > > "time":0.0}, > > "expand":{ > > "time":0.0}, > > "terms":{ > > "time":0.0}, > > "debug":{ > > "time":6.0}}, > > "loadFieldValues":{ > > "time":0.0}}}} > > > Regards, > Edwin > > > On Fri, 25 Jan 2019 at 19:52, Zheng Lin Edwin Yeo <edwinye...@gmail.com> > wrote: > >> Hi Jan and Shawn, >> >> Please focus on the strange issue that I have described above in more >> details, summary is as follows: >> >> 1. Index customers data, then queries from highlight, select, and all >> handlers are very fast (less than 50ms) >> >> 2. Now index policies data, then queries on polices are very fast BUT >> queries on customers becomes slow >> >> 3. If I reindex customers data, then again queries for customers are very >> fast BUT queries on policies becomes slow. >> >> How can you explain this behavior? >> >> We have never experienced such a strange behavior before Solr 7. >> >> Regards, >> Edwin >> >> On Fri, 25 Jan 2019 at 17:06, Zheng Lin Edwin Yeo <edwinye...@gmail.com> >> wrote: >> >>> Hi Jan, >>> >>> Referring to what you have mentioned that the highlighting takes up most >>> of the time in the first query from the policies collection, the >>> highlighting was very fast (less than 50ms) from the time it was indexed, >>> till the time after customers collection gets indexed, in which it slowed >>> down tremendously. >>> >>> Also, the slow down does not just affect on the highlight requestHandler. >>> It also affects other requestHandler like select and clustering as well. >>> All of them gets the QTime of more than 500ms. This is also proven in the >>> latest debug query that we have sent earlier, in which we have set hl=false >>> and fl=null. >>> >>> Any idea or explanation on this strange behavior? >>> Thank you for your support, as we look forward to shed some lights on >>> this issue and to resolve it. >>> >>> Regards, >>> Edwin >>> >>> On Thu, 24 Jan 2019 at 23:35, Zheng Lin Edwin Yeo <edwinye...@gmail.com> >>> wrote: >>> >>>> Hi Jan, >>>> >>>> Thanks for your reply. >>>> >>>> However, we are still getting a slow QTime of 517ms even after we set >>>> hl=false&fl=null. >>>> >>>> Below is the debug query: >>>> >>>> "debug":{ >>>> "rawquerystring":"cherry", >>>> "querystring":"cherry", >>>> "parsedquery":"searchFields_tcs:cherry", >>>> "parsedquery_toString":"searchFields_tcs:cherry", >>>> "explain":{ >>>> "46226513":"\n14.227914 = weight(searchFields_tcs:cherry in 5747763) >>>> [SchemaSimilarity], result of:\n 14.227914 = score(doc=5747763,freq=3.0 = >>>> termFreq=3.0\n), product of:\n 9.614556 = idf, computed as log(1 + >>>> (docCount - docFreq + 0.5) / (docFreq + 0.5)) from:\n 400.0 = >>>> docFreq\n 6000000.0 = docCount\n 1.4798305 = tfNorm, computed as >>>> (freq * (k1 + 1)) / (freq + k1 * (1 - b + b * fieldLength / >>>> avgFieldLength)) from:\n 3.0 = termFreq=3.0\n 1.2 = parameter >>>> k1\n 0.75 = parameter b\n 19.397041 = avgFieldLength\n 25.0 >>>> = fieldLength\n", >>>> "54088731":"\n13.937909 = weight(searchFields_tcs:cherry in 4840794) >>>> [SchemaSimilarity], result of:\n 13.937909 = score(doc=4840794,freq=3.0 = >>>> termFreq=3.0\n), product of:\n 9.614556 = idf, computed as log(1 + >>>> (docCount - docFreq + 0.5) / (docFreq + 0.5)) from:\n 400.0 = >>>> docFreq\n 6000000.0 = docCount\n 1.4496675 = tfNorm, computed as >>>> (freq * (k1 + 1)) / (freq + k1 * (1 - b + b * fieldLength / >>>> avgFieldLength)) from:\n 3.0 = termFreq=3.0\n 1.2 = parameter >>>> k1\n 0.75 = parameter b\n 19.397041 = avgFieldLength\n 27.0 >>>> = fieldLength\n", >>>> "QParser":"LuceneQParser", >>>> "timing":{ >>>> "time":517.0, >>>> "prepare":{ >>>> "time":0.0, >>>> "query":{ >>>> "time":0.0}, >>>> "facet":{ >>>> "time":0.0}, >>>> "facet_module":{ >>>> "time":0.0}, >>>> "mlt":{ >>>> "time":0.0}, >>>> "highlight":{ >>>> "time":0.0}, >>>> "stats":{ >>>> "time":0.0}, >>>> "expand":{ >>>> "time":0.0}, >>>> "terms":{ >>>> "time":0.0}, >>>> "debug":{ >>>> "time":0.0}}, >>>> "process":{ >>>> "time":516.0, >>>> "query":{ >>>> "time":15.0}, >>>> "facet":{ >>>> "time":0.0}, >>>> "facet_module":{ >>>> "time":0.0}, >>>> "mlt":{ >>>> "time":0.0}, >>>> "highlight":{ >>>> "time":0.0}, >>>> "stats":{ >>>> "time":0.0}, >>>> "expand":{ >>>> "time":0.0}, >>>> "terms":{ >>>> "time":0.0}, >>>> "debug":{ >>>> "time":500.0}}}}} >>>> >>>> Regards, >>>> Edwin >>>> >>>> >>>> On Thu, 24 Jan 2019 at 22:43, Jan Høydahl <jan....@cominvent.com> wrote: >>>> >>>>> Looks like highlighting takes most of the time on the first query >>>>> (680ms). You config seems to ask for a lot of highlighting here, like 100 >>>>> snippets of max 100000 characters etc. >>>>> Sounds to me that this might be a highlighting configuration problem. >>>>> Try to disable highlighting (hl=false) and see if you get back your speed. >>>>> Also, I see fl=* in your config, which is really asking for all fields. >>>>> Are you sure you want that, that may also be slow. Try to ask for just the >>>>> fields you will be using. >>>>> >>>>> -- >>>>> Jan Høydahl, search solution architect >>>>> Cominvent AS - www.cominvent.com >>>>> >>>>>> 24. jan. 2019 kl. 14:59 skrev Zheng Lin Edwin Yeo < >>>>> edwinye...@gmail.com>: >>>>>> >>>>>> Thanks for your reply. >>>>>> >>>>>> Below are what you have requested about our Solr setup, configurations >>>>>> files ,schema and results of debug queries: >>>>>> >>>>>> Looking forward to your advice and support on our problem. >>>>>> >>>>>> 1. System configurations >>>>>> OS: Windows 10 Pro 64 bit >>>>>> System Memory: 32GB >>>>>> CPU: Intel(R) Core(TM) i7-4790K CPU @ 4.00GHz, 4 Core(s), 8 Logical >>>>>> Processor(s) >>>>>> HDD: 3.0 TB (free 2.1 TB) SATA >>>>>> >>>>>> 2. solrconfig.xml of customers and policies collection, and solr.in >>>>> ,cmd >>>>>> which can be download from the following link: >>>>>> >>>>> https://drive.google.com/file/d/1AATjonQsEC5B0ldz27Xvx5A55Dp5ul8K/view?usp=sharing >>>>>> >>>>>> 3. The debug queries from both collections >>>>>> >>>>>> *3.1. Debug Query From Policies ( which is Slow)* >>>>>> >>>>>> "debug":{ >>>>>> >>>>>> "rawquerystring":"sherry", >>>>>> >>>>>> "querystring":"sherry", >>>>>> >>>>>> "parsedquery":"searchFields_tcs:sherry", >>>>>> >>>>>> "parsedquery_toString":"searchFields_tcs:sherry", >>>>>> >>>>>> "explain":{ >>>>>> >>>>>> "31702988":"\n14.540428 = weight(searchFields_tcs:sherry in >>>>>> 3097315) [SchemaSimilarity], result of:\n 14.540428 = >>>>>> score(doc=3097315,freq=5.0 = termFreq=5.0\n), product of:\n >>>>>> 8.907154 = idf, computed as log(1 + (docCount - docFreq + 0.5) / >>>>>> (docFreq + 0.5)) from:\n 812.0 = docFreq\n 6000000.0 = >>>>>> docCount\n 1.6324438 = tfNorm, computed as (freq * (k1 + 1)) / >>>>>> (freq + k1 * (1 - b + b * fieldLength / avgFieldLength)) from:\n >>>>>> 5.0 = termFreq=5.0\n 1.2 = parameter k1\n 0.75 = parameter >>>>>> b\n 19.397041 = avgFieldLength\n 31.0 = fieldLength\n”,.. >>>>>> >>>>>> "QParser":"LuceneQParser", >>>>>> >>>>>> "timing":{ >>>>>> >>>>>> "time":681.0, >>>>>> >>>>>> "prepare":{ >>>>>> >>>>>> "time":0.0, >>>>>> >>>>>> "query":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "facet":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "facet_module":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "mlt":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "highlight":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "stats":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "expand":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "terms":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "debug":{ >>>>>> >>>>>> "time":0.0}}, >>>>>> >>>>>> "process":{ >>>>>> >>>>>> "time":680.0, >>>>>> >>>>>> "query":{ >>>>>> >>>>>> "time":19.0}, >>>>>> >>>>>> "facet":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "facet_module":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "mlt":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "highlight":{ >>>>>> >>>>>> "time":651.0}, >>>>>> >>>>>> "stats":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "expand":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "terms":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "debug":{ >>>>>> >>>>>> "time":8.0}}, >>>>>> >>>>>> "loadFieldValues":{ >>>>>> >>>>>> "time":12.0}}}} >>>>>> >>>>>> >>>>>> >>>>>> *3.2. Debug Query From Customers (which is fast because we index it >>>>> after >>>>>> indexing Policies):* >>>>>> >>>>>> >>>>>> >>>>>> "debug":{ >>>>>> >>>>>> "rawquerystring":"sherry", >>>>>> >>>>>> "querystring":"sherry", >>>>>> >>>>>> "parsedquery":"searchFields_tcs:sherry", >>>>>> >>>>>> "parsedquery_toString":"searchFields_tcs:sherry", >>>>>> >>>>>> "explain":{ >>>>>> >>>>>> "S7900271B":"\n13.191501 = weight(searchFields_tcs:sherry in >>>>>> 2453665) [SchemaSimilarity], result of:\n 13.191501 = >>>>>> score(doc=2453665,freq=3.0 = termFreq=3.0\n), product of:\n 9.08604 >>>>>> = idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq + >>>>>> 0.5)) from:\n 428.0 = docFreq\n 3784142.0 = docCount\n >>>>>> 1.4518428 = tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 - >>>>>> b + b * fieldLength / avgFieldLength)) from:\n 3.0 = >>>>>> termFreq=3.0\n 1.2 = parameter k1\n 0.75 = parameter b\n >>>>>> 20.22558 = avgFieldLength\n 28.0 = fieldLength\n”, .. >>>>>> >>>>>> "QParser":"LuceneQParser", >>>>>> >>>>>> "timing":{ >>>>>> >>>>>> "time":38.0, >>>>>> >>>>>> "prepare":{ >>>>>> >>>>>> "time":1.0, >>>>>> >>>>>> "query":{ >>>>>> >>>>>> "time":1.0}, >>>>>> >>>>>> "facet":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "facet_module":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "mlt":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "highlight":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "stats":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "expand":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "terms":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "debug":{ >>>>>> >>>>>> "time":0.0}}, >>>>>> >>>>>> "process":{ >>>>>> >>>>>> "time":36.0, >>>>>> >>>>>> "query":{ >>>>>> >>>>>> "time":1.0}, >>>>>> >>>>>> "facet":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "facet_module":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "mlt":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "highlight":{ >>>>>> >>>>>> "time":31.0}, >>>>>> >>>>>> "stats":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "expand":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "terms":{ >>>>>> >>>>>> "time":0.0}, >>>>>> >>>>>> "debug":{ >>>>>> >>>>>> "time":3.0}}, >>>>>> >>>>>> "loadFieldValues":{ >>>>>> >>>>>> "time":13.0}}}} >>>>>> >>>>>> >>>>>> >>>>>> Best Regards, >>>>>> Edwin >>>>>> >>>>>> On Thu, 24 Jan 2019 at 20:57, Jan Høydahl <jan....@cominvent.com> >>>>> wrote: >>>>>> >>>>>>> It would be useful if you can disclose the machine configuration, OS, >>>>>>> memory, settings etc, as well as solr config including solr.in < >>>>>>> http://solr.in/>.sh, solrconfig.xml etc, so we can see the whole >>>>> picture >>>>>>> of memory, GC, etc. >>>>>>> You could also specify debugQuery=true on a slow search and check the >>>>>>> timings section for clues. What QTime are you seeing on the slow >>>>> queries in >>>>>>> solr.log? >>>>>>> If that does not reveal the reason, I'd connect to your solr >>>>> instance with >>>>>>> a tool like jVisualVM or similar, to inspect what takes time. Or >>>>> better, >>>>>>> hook up to DataDog, SPM or some other cloud tool to get a full view >>>>> of the >>>>>>> system. >>>>>>> >>>>>>> -- >>>>>>> Jan Høydahl, search solution architect >>>>>>> Cominvent AS - www.cominvent.com >>>>>>> >>>>>>>> 24. jan. 2019 kl. 13:42 skrev Zheng Lin Edwin Yeo < >>>>> edwinye...@gmail.com >>>>>>>> : >>>>>>>> >>>>>>>> Hi Shawn, >>>>>>>> >>>>>>>> Unfortunately your reply of memory may not be valid. Please refer >>>>> to my >>>>>>>> explanation below of the strange behaviors (is it much more like a >>>>> BUG >>>>>>> than >>>>>>>> anything else that is explainable): >>>>>>>> >>>>>>>> Note that we still have 18GB of free unused memory on the server. >>>>>>>> >>>>>>>> 1. We indexed the first collection called customers (3.7 millioin >>>>> records >>>>>>>> from CSV data), index size is 2.09GB. The search in customers for >>>>> any >>>>>>>> keyword is returned within 50ms (QTime) for using highlight (unified >>>>>>>> highlighter, posting, light term vectors) >>>>>>>> >>>>>>>> 2. Then we indexed the second collection called policies (6 million >>>>>>> records >>>>>>>> from CSV data), index size is 2.55GB. The search in policies for any >>>>>>>> keyword is returned within 50ms (QTime) for using highlight (unified >>>>>>>> highlighter, posting, light term vectors) >>>>>>>> >>>>>>>> 3. But now any search in customers for any keywords (not from cache) >>>>>>> takes >>>>>>>> as high as 1200ms (QTime). But still policies search remains very >>>>> fast >>>>>>>> (50ms). >>>>>>>> >>>>>>>> 4. So we decided to run the force optimize command on customers >>>>>>> collection ( >>>>>>>> >>>>>>> >>>>> https://localhost:8983/edm/customers/update?optimize=true&numSegments=1&waitFlush=false >>>>>>> ), >>>>>>>> surprisingly after optimization the search on customers collection >>>>> for >>>>>>> any >>>>>>>> keywords become very fast again (less than 50ms). BUT strangely, the >>>>>>> search >>>>>>>> in policies collection become very slow (around 1200ms) without any >>>>>>> changes >>>>>>>> to the policies collection. >>>>>>>> >>>>>>>> 5. Based on above result, we decided to run the force optimize >>>>> command on >>>>>>>> policies collection ( >>>>>>>> >>>>>>> >>>>> https://localhost:8983/edm/policies/update?optimize=true&numSegments=1&waitFlush=false >>>>>>> ). >>>>>>>> More surprisingly, after optimization the search on policies >>>>> collection >>>>>>> for >>>>>>>> any keywords become very fast again (less than 50ms). BUT more >>>>> strangely, >>>>>>>> the search in customers collection again become very slow (around >>>>> 1200ms) >>>>>>>> without any changes to the customers collection. >>>>>>>> >>>>>>>> What a strange and unexpected behavior! If this is not a bug, how >>>>> could >>>>>>> you >>>>>>>> explain the above very strange behavior in Solr 7.5. Could it be a >>>>> bug? >>>>>>>> >>>>>>>> We would appreciate any support or help on our above situation. >>>>>>>> >>>>>>>> Thank you. >>>>>>>> >>>>>>>> Regards, >>>>>>>> Edwin >>>>>>>> >>>>>>>> On Thu, 24 Jan 2019 at 16:14, Zheng Lin Edwin Yeo < >>>>> edwinye...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi Shawn, >>>>>>>>> >>>>>>>>>> If the two collections have data on the same server(s), I can see >>>>> this >>>>>>>>>> happening. More memory is consumed when there is additional >>>>> data, and >>>>>>>>>> when Solr needs more memory, performance might be affected. The >>>>>>>>>> solution is generally to install more memory in the server. >>>>>>>>> >>>>>>>>> I have found that even after we delete the index in collection2, >>>>> the >>>>>>> query >>>>>>>>> QTime for collection1 still remains slow. It does not goes back to >>>>> its >>>>>>>>> previous fast speed before we index collection2. >>>>>>>>> >>>>>>>>> Regards, >>>>>>>>> Edwin >>>>>>>>> >>>>>>>>> >>>>>>>>> On Thu, 24 Jan 2019 at 11:13, Zheng Lin Edwin Yeo < >>>>> edwinye...@gmail.com >>>>>>>> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi Shawn, >>>>>>>>>> >>>>>>>>>> Thanks for your reply. >>>>>>>>>> >>>>>>>>>> The log only shows a list the following and I don't see any >>>>> other logs >>>>>>>>>> besides these. >>>>>>>>>> >>>>>>>>>> 2019-01-24 02:47:57.925 INFO (qtp2131952342-1330) [c:collectioin1 >>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2] >>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory >>>>>>> update-script#processAdd: >>>>>>>>>> id=13245417 >>>>>>>>>> 2019-01-24 02:47:57.957 INFO (qtp2131952342-1330) [c:collectioin1 >>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2] >>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory >>>>>>> update-script#processAdd: >>>>>>>>>> id=13245430 >>>>>>>>>> 2019-01-24 02:47:57.957 INFO (qtp2131952342-1330) [c:collectioin1 >>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2] >>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory >>>>>>> update-script#processAdd: >>>>>>>>>> id=13245435 >>>>>>>>>> >>>>>>>>>> There is no change to the segments info. but the slowdown in the >>>>> first >>>>>>>>>> collection is very drastic. >>>>>>>>>> Before the indexing of collection2, the collection1 query QTime >>>>> are in >>>>>>>>>> the range of 4 to 50 ms. However, after indexing collection2, the >>>>>>>>>> collection1 query QTime increases to more than 1000 ms. The index >>>>> are >>>>>>> done >>>>>>>>>> in CSV format, and the size of the index is 3GB. >>>>>>>>>> >>>>>>>>>> Regards, >>>>>>>>>> Edwin >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Thu, 24 Jan 2019 at 01:09, Shawn Heisey <apa...@elyograg.org> >>>>>>> wrote: >>>>>>>>>> >>>>>>>>>>> On 1/23/2019 10:01 AM, Zheng Lin Edwin Yeo wrote: >>>>>>>>>>>> I am using Solr 7.5.0, and currently I am facing an issue of >>>>> when I >>>>>>> am >>>>>>>>>>>> indexing in collection2, the indexing affects the records in >>>>>>>>>>> collection1. >>>>>>>>>>>> Although the records are still intact, it seems that the >>>>> settings of >>>>>>>>>>> the >>>>>>>>>>>> termVecotrs get wipe out, and the index size of collection1 >>>>> reduced >>>>>>>>>>> from >>>>>>>>>>>> 3.3GB to 2.1GB after I do the indexing in collection2. >>>>>>>>>>> >>>>>>>>>>> This should not be possible. Indexing in one collection should >>>>> have >>>>>>>>>>> absolutely no effect on another collection. >>>>>>>>>>> >>>>>>>>>>> If logging has been left at its default settings, the solr.log >>>>> file >>>>>>>>>>> should have enough info to show what actually happened. >>>>>>>>>>> >>>>>>>>>>>> Also, the search in >>>>>>>>>>>> collection1, which was originall very fast, becomes very slow >>>>> after >>>>>>> the >>>>>>>>>>>> indexing is done is collection2. >>>>>>>>>>> >>>>>>>>>>> If the two collections have data on the same server(s), I can >>>>> see this >>>>>>>>>>> happening. More memory is consumed when there is additional >>>>> data, and >>>>>>>>>>> when Solr needs more memory, performance might be affected. The >>>>>>>>>>> solution is generally to install more memory in the server. If >>>>> the >>>>>>>>>>> system is working, there should be no need to increase the heap >>>>> size >>>>>>>>>>> when the memory size increases ... but there can be situations >>>>> where >>>>>>> the >>>>>>>>>>> heap is a little bit too small, where you WOULD want to increase >>>>> the >>>>>>>>>>> heap size. >>>>>>>>>>> >>>>>>>>>>> Thanks, >>>>>>>>>>> Shawn >>>>>>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>> >>>>> >>>>>