How do you start Solr, cause the solr.in.cmd you sent does not contain the 
memory settings. What other parameters do you start Solr with?

--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com

> 25. jan. 2019 kl. 15:28 skrev Zheng Lin Edwin Yeo <edwinye...@gmail.com>:
> 
> Hi Jan,
> 
> We are using 64 bit Java, version 1.8.0_191.
> We started Solr with 6 GB heap size.
> 
> Besides Solr, we have ZooKeeper, IIS, Google Chrome and NotePad++  running
> on the machine. There is still 22 GB of memory left on the server, out of
> the 32 GB available on the machine.
> 
> Regards,
> Edwin
> 
> On Fri, 25 Jan 2019 at 21:04, Jan Høydahl <jan....@cominvent.com> wrote:
> 
>> Which java version? 32 or 64 bit? You start Solr with default 512Mb heap
>> size?
>> Other software running on the machine?
>> 
>> --
>> Jan Høydahl, search solution architect
>> Cominvent AS - www.cominvent.com
>> 
>>> 25. jan. 2019 kl. 13:05 skrev Zheng Lin Edwin Yeo <edwinye...@gmail.com
>>> :
>>> 
>>> Hi Jan and Shawn,
>>> 
>>> For your info, this is another debug query.
>>> 
>>> "debug":{
>>> 
>>>   "rawquerystring":"johnny",
>>> 
>>>   "querystring":"johnny",
>>> 
>>>   "parsedquery":"searchFields_tcs:johnny",
>>> 
>>>   "parsedquery_toString":"searchFields_tcs:johnny",
>>> 
>>>   "explain":{
>>> 
>>>     "192280":"\n12.8497505 = weight(searchFields_tcs:johnny in
>>> 75730) [SchemaSimilarity], result of:\n  12.8497505 =
>>> score(doc=75730,freq=4.0 = termFreq=4.0\n), product of:\n    7.5108404
>>> = idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq +
>>> 0.5)) from:\n      473.0 = docFreq\n      865438.0 = docCount\n
>>> 1.7108272 = tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 -
>>> b + b * fieldLength / avgFieldLength)) from:\n      4.0 =
>>> termFreq=4.0\n      1.2 = parameter k1\n      0.75 = parameter b\n
>>> 26.66791 = avgFieldLength\n      25.0 = fieldLength\n”,
>>> 
>>>   "QParser":"LuceneQParser",
>>> 
>>>   "timing":{
>>> 
>>>     "time":350.0,
>>> 
>>>     "prepare":{
>>> 
>>>       "time":0.0,
>>> 
>>>       "query":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "facet":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "facet_module":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "mlt":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "highlight":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "stats":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "expand":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "terms":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "debug":{
>>> 
>>>         "time":0.0}},
>>> 
>>>     "process":{
>>> 
>>>       "time":348.0,
>>> 
>>>       "query":{
>>> 
>>>         "time":287.0},
>>> 
>>>       "facet":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "facet_module":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "mlt":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "highlight":{
>>> 
>>>         "time":54.0},
>>> 
>>>       "stats":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "expand":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "terms":{
>>> 
>>>         "time":0.0},
>>> 
>>>       "debug":{
>>> 
>>>         "time":6.0}},
>>> 
>>>     "loadFieldValues":{
>>> 
>>>       "time":0.0}}}}
>>> 
>>> 
>>> Regards,
>>> Edwin
>>> 
>>> 
>>> On Fri, 25 Jan 2019 at 19:52, Zheng Lin Edwin Yeo <edwinye...@gmail.com>
>>> wrote:
>>> 
>>>> Hi Jan and Shawn,
>>>> 
>>>> Please focus on the strange issue that I have described above in more
>>>> details, summary is as follows:
>>>> 
>>>> 1. Index customers data, then queries from highlight, select, and all
>>>> handlers are very fast (less than 50ms)
>>>> 
>>>> 2. Now index policies data, then queries on polices are very fast BUT
>>>> queries on customers becomes slow
>>>> 
>>>> 3. If I reindex customers data, then again queries for customers are
>> very
>>>> fast BUT queries on policies becomes slow.
>>>> 
>>>> How can you explain this behavior?
>>>> 
>>>> We have never experienced such a strange behavior before Solr 7.
>>>> 
>>>> Regards,
>>>> Edwin
>>>> 
>>>> On Fri, 25 Jan 2019 at 17:06, Zheng Lin Edwin Yeo <edwinye...@gmail.com
>>> 
>>>> wrote:
>>>> 
>>>>> Hi Jan,
>>>>> 
>>>>> Referring to what you have mentioned that the highlighting takes up
>> most
>>>>> of the time in the first query from the policies collection, the
>>>>> highlighting was very fast (less than 50ms) from the time it was
>> indexed,
>>>>> till the time after customers collection gets indexed, in which it
>> slowed
>>>>> down tremendously.
>>>>> 
>>>>> Also, the slow down does not just affect on the highlight
>> requestHandler.
>>>>> It also affects other requestHandler like select and clustering as
>> well.
>>>>> All of them gets the QTime of more than 500ms. This is also proven in
>> the
>>>>> latest debug query that we have sent earlier, in which we have set
>> hl=false
>>>>> and fl=null.
>>>>> 
>>>>> Any idea or explanation on this strange behavior?
>>>>> Thank you for your support, as we look forward to shed some lights on
>>>>> this issue and to resolve it.
>>>>> 
>>>>> Regards,
>>>>> Edwin
>>>>> 
>>>>> On Thu, 24 Jan 2019 at 23:35, Zheng Lin Edwin Yeo <
>> edwinye...@gmail.com>
>>>>> wrote:
>>>>> 
>>>>>> Hi Jan,
>>>>>> 
>>>>>> Thanks for your reply.
>>>>>> 
>>>>>> However, we are still getting a slow QTime of 517ms even after we set
>>>>>> hl=false&fl=null.
>>>>>> 
>>>>>> Below is the debug query:
>>>>>> 
>>>>>> "debug":{
>>>>>>   "rawquerystring":"cherry",
>>>>>>   "querystring":"cherry",
>>>>>>   "parsedquery":"searchFields_tcs:cherry",
>>>>>>   "parsedquery_toString":"searchFields_tcs:cherry",
>>>>>>   "explain":{
>>>>>>     "46226513":"\n14.227914 = weight(searchFields_tcs:cherry in
>> 5747763) [SchemaSimilarity], result of:\n  14.227914 =
>> score(doc=5747763,freq=3.0 = termFreq=3.0\n), product of:\n    9.614556 =
>> idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq + 0.5))
>> from:\n      400.0 = docFreq\n      6000000.0 = docCount\n    1.4798305 =
>> tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 - b + b *
>> fieldLength / avgFieldLength)) from:\n      3.0 = termFreq=3.0\n      1.2 =
>> parameter k1\n      0.75 = parameter b\n      19.397041 = avgFieldLength\n
>>    25.0 = fieldLength\n",
>>>>>>     "54088731":"\n13.937909 = weight(searchFields_tcs:cherry in
>> 4840794) [SchemaSimilarity], result of:\n  13.937909 =
>> score(doc=4840794,freq=3.0 = termFreq=3.0\n), product of:\n    9.614556 =
>> idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq + 0.5))
>> from:\n      400.0 = docFreq\n      6000000.0 = docCount\n    1.4496675 =
>> tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1 - b + b *
>> fieldLength / avgFieldLength)) from:\n      3.0 = termFreq=3.0\n      1.2 =
>> parameter k1\n      0.75 = parameter b\n      19.397041 = avgFieldLength\n
>>    27.0 = fieldLength\n",
>>>>>>   "QParser":"LuceneQParser",
>>>>>>   "timing":{
>>>>>>     "time":517.0,
>>>>>>     "prepare":{
>>>>>>       "time":0.0,
>>>>>>       "query":{
>>>>>>         "time":0.0},
>>>>>>       "facet":{
>>>>>>         "time":0.0},
>>>>>>       "facet_module":{
>>>>>>         "time":0.0},
>>>>>>       "mlt":{
>>>>>>         "time":0.0},
>>>>>>       "highlight":{
>>>>>>         "time":0.0},
>>>>>>       "stats":{
>>>>>>         "time":0.0},
>>>>>>       "expand":{
>>>>>>         "time":0.0},
>>>>>>       "terms":{
>>>>>>         "time":0.0},
>>>>>>       "debug":{
>>>>>>         "time":0.0}},
>>>>>>     "process":{
>>>>>>       "time":516.0,
>>>>>>       "query":{
>>>>>>         "time":15.0},
>>>>>>       "facet":{
>>>>>>         "time":0.0},
>>>>>>       "facet_module":{
>>>>>>         "time":0.0},
>>>>>>       "mlt":{
>>>>>>         "time":0.0},
>>>>>>       "highlight":{
>>>>>>         "time":0.0},
>>>>>>       "stats":{
>>>>>>         "time":0.0},
>>>>>>       "expand":{
>>>>>>         "time":0.0},
>>>>>>       "terms":{
>>>>>>         "time":0.0},
>>>>>>       "debug":{
>>>>>>         "time":500.0}}}}}
>>>>>> 
>>>>>> Regards,
>>>>>> Edwin
>>>>>> 
>>>>>> 
>>>>>> On Thu, 24 Jan 2019 at 22:43, Jan Høydahl <jan....@cominvent.com>
>> wrote:
>>>>>> 
>>>>>>> Looks like highlighting takes most of the time on the first query
>>>>>>> (680ms). You config seems to ask for a lot of highlighting here,
>> like 100
>>>>>>> snippets of max 100000 characters etc.
>>>>>>> Sounds to me that this might be a highlighting configuration problem.
>>>>>>> Try to disable highlighting (hl=false) and see if you get back your
>> speed.
>>>>>>> Also, I see fl=* in your config, which is really asking for all
>> fields.
>>>>>>> Are you sure you want that, that may also be slow. Try to ask for
>> just the
>>>>>>> fields you will be using.
>>>>>>> 
>>>>>>> --
>>>>>>> Jan Høydahl, search solution architect
>>>>>>> Cominvent AS - www.cominvent.com
>>>>>>> 
>>>>>>>> 24. jan. 2019 kl. 14:59 skrev Zheng Lin Edwin Yeo <
>>>>>>> edwinye...@gmail.com>:
>>>>>>>> 
>>>>>>>> Thanks for your reply.
>>>>>>>> 
>>>>>>>> Below are what you have requested about our Solr setup,
>> configurations
>>>>>>>> files ,schema and results of debug queries:
>>>>>>>> 
>>>>>>>> Looking forward to your advice and support on our problem.
>>>>>>>> 
>>>>>>>> 1. System configurations
>>>>>>>> OS: Windows 10 Pro 64 bit
>>>>>>>> System Memory: 32GB
>>>>>>>> CPU: Intel(R) Core(TM) i7-4790K CPU @ 4.00GHz, 4 Core(s), 8 Logical
>>>>>>>> Processor(s)
>>>>>>>> HDD: 3.0 TB (free 2.1 TB)  SATA
>>>>>>>> 
>>>>>>>> 2. solrconfig.xml of customers and policies collection, and solr.in
>>>>>>> ,cmd
>>>>>>>> which can be download from the following link:
>>>>>>>> 
>>>>>>> 
>> https://drive.google.com/file/d/1AATjonQsEC5B0ldz27Xvx5A55Dp5ul8K/view?usp=sharing
>>>>>>>> 
>>>>>>>> 3. The debug queries from both collections
>>>>>>>> 
>>>>>>>> *3.1. Debug Query From Policies ( which is Slow)*
>>>>>>>> 
>>>>>>>> "debug":{
>>>>>>>> 
>>>>>>>>  "rawquerystring":"sherry",
>>>>>>>> 
>>>>>>>>  "querystring":"sherry",
>>>>>>>> 
>>>>>>>>  "parsedquery":"searchFields_tcs:sherry",
>>>>>>>> 
>>>>>>>>  "parsedquery_toString":"searchFields_tcs:sherry",
>>>>>>>> 
>>>>>>>>  "explain":{
>>>>>>>> 
>>>>>>>>    "31702988":"\n14.540428 = weight(searchFields_tcs:sherry in
>>>>>>>> 3097315) [SchemaSimilarity], result of:\n  14.540428 =
>>>>>>>> score(doc=3097315,freq=5.0 = termFreq=5.0\n), product of:\n
>>>>>>>> 8.907154 = idf, computed as log(1 + (docCount - docFreq + 0.5) /
>>>>>>>> (docFreq + 0.5)) from:\n      812.0 = docFreq\n      6000000.0 =
>>>>>>>> docCount\n    1.6324438 = tfNorm, computed as (freq * (k1 + 1)) /
>>>>>>>> (freq + k1 * (1 - b + b * fieldLength / avgFieldLength)) from:\n
>>>>>>>> 5.0 = termFreq=5.0\n      1.2 = parameter k1\n      0.75 = parameter
>>>>>>>> b\n      19.397041 = avgFieldLength\n      31.0 = fieldLength\n”,..
>>>>>>>> 
>>>>>>>>  "QParser":"LuceneQParser",
>>>>>>>> 
>>>>>>>>  "timing":{
>>>>>>>> 
>>>>>>>>    "time":681.0,
>>>>>>>> 
>>>>>>>>    "prepare":{
>>>>>>>> 
>>>>>>>>      "time":0.0,
>>>>>>>> 
>>>>>>>>      "query":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "facet":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "facet_module":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "mlt":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "highlight":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "stats":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "expand":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "terms":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "debug":{
>>>>>>>> 
>>>>>>>>        "time":0.0}},
>>>>>>>> 
>>>>>>>>    "process":{
>>>>>>>> 
>>>>>>>>      "time":680.0,
>>>>>>>> 
>>>>>>>>      "query":{
>>>>>>>> 
>>>>>>>>        "time":19.0},
>>>>>>>> 
>>>>>>>>      "facet":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "facet_module":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "mlt":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "highlight":{
>>>>>>>> 
>>>>>>>>        "time":651.0},
>>>>>>>> 
>>>>>>>>      "stats":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "expand":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "terms":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "debug":{
>>>>>>>> 
>>>>>>>>        "time":8.0}},
>>>>>>>> 
>>>>>>>>    "loadFieldValues":{
>>>>>>>> 
>>>>>>>>      "time":12.0}}}}
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> *3.2. Debug Query From Customers (which is fast because we index it
>>>>>>> after
>>>>>>>> indexing Policies):*
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> "debug":{
>>>>>>>> 
>>>>>>>>  "rawquerystring":"sherry",
>>>>>>>> 
>>>>>>>>  "querystring":"sherry",
>>>>>>>> 
>>>>>>>>  "parsedquery":"searchFields_tcs:sherry",
>>>>>>>> 
>>>>>>>>  "parsedquery_toString":"searchFields_tcs:sherry",
>>>>>>>> 
>>>>>>>>  "explain":{
>>>>>>>> 
>>>>>>>>    "S7900271B":"\n13.191501 = weight(searchFields_tcs:sherry in
>>>>>>>> 2453665) [SchemaSimilarity], result of:\n  13.191501 =
>>>>>>>> score(doc=2453665,freq=3.0 = termFreq=3.0\n), product of:\n
>> 9.08604
>>>>>>>> = idf, computed as log(1 + (docCount - docFreq + 0.5) / (docFreq +
>>>>>>>> 0.5)) from:\n      428.0 = docFreq\n      3784142.0 = docCount\n
>>>>>>>> 1.4518428 = tfNorm, computed as (freq * (k1 + 1)) / (freq + k1 * (1
>> -
>>>>>>>> b + b * fieldLength / avgFieldLength)) from:\n      3.0 =
>>>>>>>> termFreq=3.0\n      1.2 = parameter k1\n      0.75 = parameter b\n
>>>>>>>> 20.22558 = avgFieldLength\n      28.0 = fieldLength\n”, ..
>>>>>>>> 
>>>>>>>>  "QParser":"LuceneQParser",
>>>>>>>> 
>>>>>>>>  "timing":{
>>>>>>>> 
>>>>>>>>    "time":38.0,
>>>>>>>> 
>>>>>>>>    "prepare":{
>>>>>>>> 
>>>>>>>>      "time":1.0,
>>>>>>>> 
>>>>>>>>      "query":{
>>>>>>>> 
>>>>>>>>        "time":1.0},
>>>>>>>> 
>>>>>>>>      "facet":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "facet_module":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "mlt":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "highlight":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "stats":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "expand":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "terms":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "debug":{
>>>>>>>> 
>>>>>>>>        "time":0.0}},
>>>>>>>> 
>>>>>>>>    "process":{
>>>>>>>> 
>>>>>>>>      "time":36.0,
>>>>>>>> 
>>>>>>>>      "query":{
>>>>>>>> 
>>>>>>>>        "time":1.0},
>>>>>>>> 
>>>>>>>>      "facet":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "facet_module":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "mlt":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "highlight":{
>>>>>>>> 
>>>>>>>>        "time":31.0},
>>>>>>>> 
>>>>>>>>      "stats":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "expand":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "terms":{
>>>>>>>> 
>>>>>>>>        "time":0.0},
>>>>>>>> 
>>>>>>>>      "debug":{
>>>>>>>> 
>>>>>>>>        "time":3.0}},
>>>>>>>> 
>>>>>>>>    "loadFieldValues":{
>>>>>>>> 
>>>>>>>>      "time":13.0}}}}
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> Best Regards,
>>>>>>>> Edwin
>>>>>>>> 
>>>>>>>> On Thu, 24 Jan 2019 at 20:57, Jan Høydahl <jan....@cominvent.com>
>>>>>>> wrote:
>>>>>>>> 
>>>>>>>>> It would be useful if you can disclose the machine configuration,
>> OS,
>>>>>>>>> memory, settings etc, as well as solr config including solr.in <
>>>>>>>>> http://solr.in/>.sh, solrconfig.xml etc, so we can see the whole
>>>>>>> picture
>>>>>>>>> of memory, GC, etc.
>>>>>>>>> You could also specify debugQuery=true on a slow search and check
>> the
>>>>>>>>> timings section for clues. What QTime are you seeing on the slow
>>>>>>> queries in
>>>>>>>>> solr.log?
>>>>>>>>> If that does not reveal the reason, I'd connect to your solr
>>>>>>> instance with
>>>>>>>>> a tool like jVisualVM or similar, to inspect what takes time. Or
>>>>>>> better,
>>>>>>>>> hook up to DataDog, SPM or some other cloud tool to get a full view
>>>>>>> of the
>>>>>>>>> system.
>>>>>>>>> 
>>>>>>>>> --
>>>>>>>>> Jan Høydahl, search solution architect
>>>>>>>>> Cominvent AS - www.cominvent.com
>>>>>>>>> 
>>>>>>>>>> 24. jan. 2019 kl. 13:42 skrev Zheng Lin Edwin Yeo <
>>>>>>> edwinye...@gmail.com
>>>>>>>>>> :
>>>>>>>>>> 
>>>>>>>>>> Hi Shawn,
>>>>>>>>>> 
>>>>>>>>>> Unfortunately your reply of memory may not be valid. Please refer
>>>>>>> to my
>>>>>>>>>> explanation below of the strange behaviors (is it much more like a
>>>>>>> BUG
>>>>>>>>> than
>>>>>>>>>> anything else that is explainable):
>>>>>>>>>> 
>>>>>>>>>> Note that we still have 18GB of free unused memory on the server.
>>>>>>>>>> 
>>>>>>>>>> 1. We indexed the first collection called customers (3.7 millioin
>>>>>>> records
>>>>>>>>>> from CSV data), index size is 2.09GB. The search in customers for
>>>>>>> any
>>>>>>>>>> keyword is returned within 50ms (QTime) for using highlight
>> (unified
>>>>>>>>>> highlighter, posting, light term vectors)
>>>>>>>>>> 
>>>>>>>>>> 2. Then we indexed the second collection called policies (6
>> million
>>>>>>>>> records
>>>>>>>>>> from CSV data), index size is 2.55GB. The search in policies for
>> any
>>>>>>>>>> keyword is returned within 50ms (QTime) for using highlight
>> (unified
>>>>>>>>>> highlighter, posting, light term vectors)
>>>>>>>>>> 
>>>>>>>>>> 3. But now any search in customers for any keywords (not from
>> cache)
>>>>>>>>> takes
>>>>>>>>>> as high as 1200ms (QTime). But still policies search remains very
>>>>>>> fast
>>>>>>>>>> (50ms).
>>>>>>>>>> 
>>>>>>>>>> 4. So we decided to run the force optimize command on customers
>>>>>>>>> collection (
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>> 
>> https://localhost:8983/edm/customers/update?optimize=true&numSegments=1&waitFlush=false
>>>>>>>>> ),
>>>>>>>>>> surprisingly after optimization the search on customers collection
>>>>>>> for
>>>>>>>>> any
>>>>>>>>>> keywords become very fast again (less than 50ms). BUT strangely,
>> the
>>>>>>>>> search
>>>>>>>>>> in policies collection become very slow (around 1200ms) without
>> any
>>>>>>>>> changes
>>>>>>>>>> to the policies collection.
>>>>>>>>>> 
>>>>>>>>>> 5. Based on above result, we decided to run the force optimize
>>>>>>> command on
>>>>>>>>>> policies collection (
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>> 
>> https://localhost:8983/edm/policies/update?optimize=true&numSegments=1&waitFlush=false
>>>>>>>>> ).
>>>>>>>>>> More surprisingly, after optimization the search on policies
>>>>>>> collection
>>>>>>>>> for
>>>>>>>>>> any keywords become very fast again (less than 50ms). BUT more
>>>>>>> strangely,
>>>>>>>>>> the search in customers collection again become very slow (around
>>>>>>> 1200ms)
>>>>>>>>>> without any changes to the customers collection.
>>>>>>>>>> 
>>>>>>>>>> What a strange and unexpected behavior! If this is not a bug, how
>>>>>>> could
>>>>>>>>> you
>>>>>>>>>> explain the above very strange behavior in Solr 7.5. Could it be a
>>>>>>> bug?
>>>>>>>>>> 
>>>>>>>>>> We would appreciate any support or help on our above situation.
>>>>>>>>>> 
>>>>>>>>>> Thank you.
>>>>>>>>>> 
>>>>>>>>>> Regards,
>>>>>>>>>> Edwin
>>>>>>>>>> 
>>>>>>>>>> On Thu, 24 Jan 2019 at 16:14, Zheng Lin Edwin Yeo <
>>>>>>> edwinye...@gmail.com>
>>>>>>>>>> wrote:
>>>>>>>>>> 
>>>>>>>>>>> Hi Shawn,
>>>>>>>>>>> 
>>>>>>>>>>>> If the two collections have data on the same server(s), I can
>> see
>>>>>>> this
>>>>>>>>>>>> happening.  More memory is consumed when there is additional
>>>>>>> data, and
>>>>>>>>>>>> when Solr needs more memory, performance might be affected.  The
>>>>>>>>>>>> solution is generally to install more memory in the server.
>>>>>>>>>>> 
>>>>>>>>>>> I have found that even after we delete the index in collection2,
>>>>>>> the
>>>>>>>>> query
>>>>>>>>>>> QTime for collection1 still remains slow. It does not goes back
>> to
>>>>>>> its
>>>>>>>>>>> previous fast speed before we index collection2.
>>>>>>>>>>> 
>>>>>>>>>>> Regards,
>>>>>>>>>>> Edwin
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> On Thu, 24 Jan 2019 at 11:13, Zheng Lin Edwin Yeo <
>>>>>>> edwinye...@gmail.com
>>>>>>>>>> 
>>>>>>>>>>> wrote:
>>>>>>>>>>> 
>>>>>>>>>>>> Hi Shawn,
>>>>>>>>>>>> 
>>>>>>>>>>>> Thanks for your reply.
>>>>>>>>>>>> 
>>>>>>>>>>>> The log only shows a list  the following and I don't see any
>>>>>>> other logs
>>>>>>>>>>>> besides these.
>>>>>>>>>>>> 
>>>>>>>>>>>> 2019-01-24 02:47:57.925 INFO  (qtp2131952342-1330)
>> [c:collectioin1
>>>>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2]
>>>>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory
>>>>>>>>> update-script#processAdd:
>>>>>>>>>>>> id=13245417
>>>>>>>>>>>> 2019-01-24 02:47:57.957 INFO  (qtp2131952342-1330)
>> [c:collectioin1
>>>>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2]
>>>>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory
>>>>>>>>> update-script#processAdd:
>>>>>>>>>>>> id=13245430
>>>>>>>>>>>> 2019-01-24 02:47:57.957 INFO  (qtp2131952342-1330)
>> [c:collectioin1
>>>>>>>>>>>> s:shard1 r:core_node4 x:collection1_shard1_replica_n2]
>>>>>>>>>>>> o.a.s.u.p.StatelessScriptUpdateProcessorFactory
>>>>>>>>> update-script#processAdd:
>>>>>>>>>>>> id=13245435
>>>>>>>>>>>> 
>>>>>>>>>>>> There is no change to the segments info. but the slowdown in the
>>>>>>> first
>>>>>>>>>>>> collection is very drastic.
>>>>>>>>>>>> Before the indexing of collection2, the collection1 query QTime
>>>>>>> are in
>>>>>>>>>>>> the range of 4 to 50 ms. However, after indexing collection2,
>> the
>>>>>>>>>>>> collection1 query QTime increases to more than 1000 ms. The
>> index
>>>>>>> are
>>>>>>>>> done
>>>>>>>>>>>> in CSV format, and the size of the index is 3GB.
>>>>>>>>>>>> 
>>>>>>>>>>>> Regards,
>>>>>>>>>>>> Edwin
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> On Thu, 24 Jan 2019 at 01:09, Shawn Heisey <apa...@elyograg.org
>>> 
>>>>>>>>> wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>>> On 1/23/2019 10:01 AM, Zheng Lin Edwin Yeo wrote:
>>>>>>>>>>>>>> I am using Solr 7.5.0, and currently I am facing an issue of
>>>>>>> when I
>>>>>>>>> am
>>>>>>>>>>>>>> indexing in collection2, the indexing affects the records in
>>>>>>>>>>>>> collection1.
>>>>>>>>>>>>>> Although the records are still intact, it seems that the
>>>>>>> settings of
>>>>>>>>>>>>> the
>>>>>>>>>>>>>> termVecotrs get wipe out, and the index size of collection1
>>>>>>> reduced
>>>>>>>>>>>>> from
>>>>>>>>>>>>>> 3.3GB to 2.1GB after I do the indexing in collection2.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> This should not be possible.  Indexing in one collection should
>>>>>>> have
>>>>>>>>>>>>> absolutely no effect on another collection.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> If logging has been left at its default settings, the solr.log
>>>>>>> file
>>>>>>>>>>>>> should have enough info to show what actually happened.
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Also, the search in
>>>>>>>>>>>>>> collection1, which was originall very fast, becomes very slow
>>>>>>> after
>>>>>>>>> the
>>>>>>>>>>>>>> indexing is done is collection2.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> If the two collections have data on the same server(s), I can
>>>>>>> see this
>>>>>>>>>>>>> happening.  More memory is consumed when there is additional
>>>>>>> data, and
>>>>>>>>>>>>> when Solr needs more memory, performance might be affected.
>> The
>>>>>>>>>>>>> solution is generally to install more memory in the server.  If
>>>>>>> the
>>>>>>>>>>>>> system is working, there should be no need to increase the heap
>>>>>>> size
>>>>>>>>>>>>> when the memory size increases ... but there can be situations
>>>>>>> where
>>>>>>>>> the
>>>>>>>>>>>>> heap is a little bit too small, where you WOULD want to
>> increase
>>>>>>> the
>>>>>>>>>>>>> heap size.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> Shawn
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>> 
>>>>>>> 
>> 
>> 

Reply via email to