Re: How to stop backup once initiated
There is no way to stop it once initiated. On Fri, Mar 21, 2014 at 8:09 PM, search engn dev sachinyadav0...@gmail.com wrote: My index size 20 GB and I have issues solr backup command , now this backup is going on its taking too much time , so how can i stop backup command? -- View this message in context: http://lucene.472066.n3.nabble.com/How-to-stop-backup-once-initiated-tp4126020.html Sent from the Solr - User mailing list archive at Nabble.com. -- Regards, Shalin Shekhar Mangar.
Re: setting up solr on tomcat
On 3/22/2014 2:16 AM, anupamk wrote: Hi, Is the solrTomcat wiki article valid for solr-4.7.0 ? http://wiki.apache.org/solr/SolrTomcat I am not able to deploy solr after following the instructions there. When I try to access the solr admin page I get a 404. I followed every step exactly as mentioned in the wiki, still no dice. Any ideas ? -- View this message in context: http://lucene.472066.n3.nabble.com/setting-up-solr-on-tomcat-tp4126177.html Sent from the Solr - User mailing list archive at Nabble.com. There was a note on that page saying: Solr4.3 http://wiki.apache.org/solr/Solr4.3requires completely different deployment. These instructions are*not*current and are for an indeterminate version of Solr. I haven't read the instructions in detail, but in my experience setting up a single standalone server goes like this: copy solr.war to the tomcat/webapps folder, logging jars (log4j, slf4j) and configuration (log4j.properties) to the tomcat/lib folder you can create your solr home directory directly in the tomcat folder -- if you do that, it should be found, or you can put it somewhere else and start the jvm with -Dsolr.solr.home=/wherever/you/put/solr that's pretty much it, I think. You will see the solr admin at http://localhost:8080/solr if you use all vanilla settings. -Mike
Re: understand debuginfo from query
How are you getting the impression that is not worth anything? It appears to be worth more than the inhaltstyp field, but less than the title field - as you have specified by your field boosts. -- Jack Krupansky -Original Message- From: ao...@hispeed.ch Sent: Friday, March 21, 2014 4:59 AM To: solr-user@lucene.apache.org Subject: Re: understand debuginfo from query i found a good page to explain the debug output but it is still unclear for me. why is the field plain_text not worth anything? the query term was found 3 times. you can see it here: http://explain.solr.pl/explains/a90aze3o ao...@hispeed.ch schrieb: i want the infos simplified so that the user can see why a doc was found bellow is the output a a doc: 0.085597195 = (MATCH) sum of: 0.083729245 = (MATCH) max of: 0.0019158133 = (MATCH) weight(plain_text:test^10.0 in 601) [DefaultSimilarity], result of: 0.0019158133 = score(doc=601,freq=9.0 = termFreq=9.0 ), product of: 0.022560213 = queryWeight, product of: 10.0 = boost 3.6232536 = idf(docFreq=81, maxDocs=1130) 6.2265067E-4 = queryNorm 0.084920004 = fieldWeight in 601, product of: 3.0 = tf(freq=9.0), with freq of: 9.0 = termFreq=9.0 3.6232536 = idf(docFreq=81, maxDocs=1130) 0.0078125 = fieldNorm(doc=601) 0.083729245 = (MATCH) weight(inhaltstyp:test^6.0 in 601) [DefaultSimilarity], result of: 0.083729245 = score(doc=601,freq=1.0 = termFreq=1.0 ), product of: 0.017686278 = queryWeight, product of: 6.0 = boost 4.734136 = idf(docFreq=26, maxDocs=1130) 6.2265067E-4 = queryNorm 4.734136 = fieldWeight in 601, product of: 1.0 = tf(freq=1.0), with freq of: 1.0 = termFreq=1.0 4.734136 = idf(docFreq=26, maxDocs=1130) 1.0 = fieldNorm(doc=601) 0.013458222 = (MATCH) weight(title:test^20.0 in 601) [DefaultSimilarity], result of: 0.013458222 = score(doc=601,freq=1.0 = termFreq=1.0 ), product of: 0.042281017 = queryWeight, product of: 20.0 = boost 3.395244 = idf(docFreq=102, maxDocs=1130) 6.2265067E-4 = queryNorm 0.31830412 = fieldWeight in 601, product of: 1.0 = tf(freq=1.0), with freq of: 1.0 = termFreq=1.0 3.395244 = idf(docFreq=102, maxDocs=1130) 0.09375 = fieldNorm(doc=601) 0.001867952 = (MATCH) product of: 0.003735904 = (MATCH) sum of: 0.003735904 = (MATCH) ConstantScore(expiration:[1395328539325 TO *]), product of: 1.0 = boost 0.003735904 = queryNorm 0.5 = coord(1/2) 0.0 = (MATCH) FunctionQuery(div(int(clicks),max(int(displays),const(1, product of: 0.0 = div(int(clicks)=0,max(int(displays)=432,const(1))) 8.0 = boost 6.2265067E-4 = queryNorm why is the sum 0.085597195? this would mean 0.083729245 + 0.001867952 and these are not included in the sum: 0.0019158133 + 0.013458222 + 0.003735904 am i looking at the wrong total? aren't these 2 cases the ones i have to sum up x = (MATCH) sum of or x = score( ? i'm trying to extract the fields that where used for weighing the doc.
SolrCloud from Stopping recovery for warnings to crash
Solr version: 4.7 Architecture: 2 solrs (1 shard, leader + replica) 3 zookeepers Servers: * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores * zookeeper Solr data: * 21 collections * Many fields, small docs, docs count per collection from 1k to 500k About a week ago solr started crashing. It crashes every day, 3-4 times a day. Usually at nigh. I can't tell anything what could it be related to because at that time we haven't done any configuration changes. Load haven't changed too. Everything starts with Stopping recovery for .. warnings (every warnings is repeated several times): WARN org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for zkNodeName=core_node1core=** WARN org.apache.solr.cloud.ElectionContext; cancelElection did not find election node to remove WARN org.apache.solr.update.PeerSync; no frame of reference to tell if we've missed updates WARN - 2014-03-23 04:00:26.286; org.apache.solr.update.PeerSync; no frame of reference to tell if we've missed updates WARN - 2014-03-23 04:00:30.728; org.apache.solr.handler.SnapPuller; File _f9m_Lucene41_0.doc expected to be 6218278 while it is 7759879 WARN - 2014-03-23 04:00:54.126; org.apache.solr.update.UpdateLog$LogReplayer; Starting log replay tlog{file=/path/solr/collection1_shard1_replica2/data/tlog/tlog.0003272 refcount=2} active=true starting pos=356216606 Then again Stopping recovery for .. warnings: WARN org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for zkNodeName=core_node1core=** ERROR - 2014-03-23 05:19:29.566; org.apache.solr.common.SolrException; org.apache.solr.common.SolrException: No registered leader was found after waiting for 4000ms , collection: collection1 slice: shard1 ERROR - 2014-03-23 05:20:03.961; org.apache.solr.common.SolrException; org.apache.solr.common.SolrException: I was asked to wait on state down for IP:PORT_solr but I still do not see the requested state. I see state: active live:false After this serves mostly didn't recover.
Re: Shingles in solr for bigrams,trigrams in parsed_query
The query parser only presents the query terms one at a time to the analyzer, so your analyzer doesn't see both terms on one analysis call. If you enclose your query terms in quotes as a single phrase, you should see multiple terms being processed. q=best price -- Jack Krupansky -Original Message- From: Jyotirmoy Sundi Sent: Thursday, March 20, 2014 2:06 PM To: solr-user@lucene.apache.org Subject: Singles in solr for bigrams,trigrams in parsed_query Hi Folks, I am using singles to index bigrams/trigrams. The same is also used for query in the schema.xml file. But when I run the query in debug mode for a collections, I dont see the bigrams in the parsed_query . Any idea what I might be missing. solr/colection/select?q=best%20pricedebugQuery=on str name=parsedquery_toStringtext:best text:price/str I was hoping to see str name=parsedquery_toStringtext:best text:price text:best price/str My schema files looks like this: types fieldType name=string class=solr.StrField sortMissingLast=true omitNorms=true/ fieldType name=int class=solr.TrieIntField precisionStep=0 omitNorms=true positionIncrementGap=0/ fieldType name=text class=solr.TextField positionIncrementGap=100 analyzer type=index charFilter class=solr.HTMLStripCharFilterFactory/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=4 outputUnigrams=true / tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.LengthFilterFactory min=3 max=50 / filter class=solr.WordDelimiterFilterFactory generateWordParts=0 generateNumberParts=0 catenateWords=1 catenateNumbers=1 catenateAll=1 preserveOriginal=1 splitOnCaseChange=0 splitOnNumerics=0 stemEnglishPossessive=1/ filter class=solr.StopFilterFactory/ filter class=solr.SynonymFilterFactory synonyms=synonyms.txt ignoreCase=true expand=true/ filter class=solr.TrimFilterFactory / /analyzer analyzer type=query filter class=solr.LowerCaseFilterFactory/ filter class=solr.LengthFilterFactory min=3 max=50 / tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.StopFilterFactory/ filter class=solr.TrimFilterFactory / filter class=solr.SynonymFilterFactory synonyms=synonyms.txt ignoreCase=true expand=true/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=1 catenateNumbers=1 catenateAll=1 splitOnCaseChange=0 splitOnNumerics=0 stemEnglishPossessive=1/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=4 outputUnigrams=true / filter class=solr.CommonGramsFilterFactory words=stopwords.txt ignoreCase=true/ !--filter class=solr.CommonGramsFilterFactory words=stopwords.txt ignoreCase=true/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=4 outputUnigrams=true /-- /analyzer /fieldType /types -- Best Regards, Jyotirmoy Sundi
Re: Multilingual indexing, search results, edismax and stopwords
Setting the default query operator to AND is the preferred approach: q.op=AND. That said, I'm not sure that counting ignored and empty terms towards the mm % makes sense. IOW, if a term transforms to nothing, either because it is a stop word or empty synonym replacement or pure punctuation, I don't think it should count as a term. I think this is worth a Jira. -- Jack Krupansky -Original Message- From: kastania44 Sent: Thursday, March 20, 2014 11:00 AM To: solr-user@lucene.apache.org Subject: Multilingual indexing, search results, edismax and stopwords On our drupal multilingual system we use apache Solr 3.5. The problem is well known on different blogs, sites I read. The search results are not the one we want. On our code in hook apachesolr_query_alter we override the defaultOperator: $query-replaceParam('mm', '90%'); The requirement is, when I search for: biological analyses, I want to fetch only the results which have both of the words. When I search for: biological and chemical analyses, I want it to fetch only the results which have biological , chemical, analyses. The and is not indexed due to stopwords. If I set mm to 100% and my query has stopwords it will not fetch any result. If I set mm to 100$ and my query does not have stopwords it will fetch the desired results. If I set mm anything between 50%-99% it fetches not wanted results, as results that contain only one of the searched keywords, or words like the searched keywords, like analyse (even if I searched for analyses). If I search using + before the words that are mandatory it works ok, but it is not user friently, to ask from the user to type + before each word exvcept from the stopwords. Do I make any sense? Below are some of our configuration details: All the indexed fields are of type text_language, e.g from our schema.xml /field name=label type=text indexed=true stored=true termVectors=true omitNorms=true/ field name=i18n_label_en type=text_en indexed=true stored=true termVectors=true omitNorms=true/ field name=i18n_label_fr type=text_fr indexed=true stored=true termVectors=true omitNorms=true// All the text fieldtypes have the same configuration except from the protected, words, dictionary parameters which are language specific. e.g from our schema.xml /fieldType name=text_en class=solr.TextField positionIncrementGap=100 analyzer type=index charFilter class=solr.MappingCharFilterFactory mapping=mapping-ISOLatin1Accent_en.txt/ tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.StopFilterFactory ignoreCase=true words=stopwords_en.txt enablePositionIncrements=true/ filter class=solr.WordDelimiterFilterFactory protected=protwords.txt generateWordParts=1 generateNumberParts=1 catenateWords=1 catenateNumbers=1 catenateAll=1 splitOnCaseChange=1 preserveOriginal=1 splitOnNumerics=1 stemEnglishPossessive=1/ filter class=solr.LengthFilterFactory min=2 max=100/ filter class=solr.LowerCaseFilterFactory/filter class=solr.DictionaryCompoundWordTokenFilterFactory dictionary=compoundwords_en.txt minWordSize=5 minSubwordSize=4 maxSubwordSize=15 onlyLongestMatch=true/ filter class=solr.SnowballPorterFilterFactory language=English protected=protwords_en.txt/ filter class=solr.RemoveDuplicatesTokenFilterFactory/ /analyzer analyzer type=query charFilter class=solr.MappingCharFilterFactory mapping=mapping-ISOLatin1Accent_en.txt/ tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.SynonymFilterFactory synonyms=synonyms_en.txt ignoreCase=true expand=true/ filter class=solr.StopFilterFactory ignoreCase=true words=stopwords_en.txt enablePositionIncrements=true/ filter class=solr.WordDelimiterFilterFactory protected=protwords.txt generateWordParts=1 generateNumberParts=1 catenateWords=1 catenateNumbers=1 catenateAll=1 splitOnCaseChange=1 preserveOriginal=1 splitOnNumerics=1 stemEnglishPossessive=1/ filter class=solr.LengthFilterFactory min=2 max=100/ filter class=solr.LowerCaseFilterFactory/filter class=solr.DictionaryCompoundWordTokenFilterFactory dictionary=compoundwords_en.txt minWordSize=5 minSubwordSize=4 maxSubwordSize=15 onlyLongestMatch=true/ filter class=solr.SnowballPorterFilterFactory language=English protected=protwords_en.txt/ filter class=solr.RemoveDuplicatesTokenFilterFactory/ /analyzer /fieldType/ solrQueryParser defaultOperator=AND/ solrconfig.xml /requestHandler name=pinkPony class=solr.SearchHandler default=true lst name=defaults str name=defTypeedismax/str str name=echoParamsexplicit/str bool name=omitHeadertrue/bool float name=tie0.01/float int name=timeAllowed${solr.pinkPony.timeAllowed:-1}/int str name=q.alt*:*/str str name=spellcheckfalse/str str name=spellcheck.onlyMorePopulartrue/str str name=spellcheck.extendedResultsfalse/str str name=spellcheck.count1/str /lst arr
Re: Shingles in solr for bigrams,trigrams in parsed_query
Hi Jack, Thanks for your response, but if I try q=best quality and best price, the parsedquery comes as following which is a lot of unwanted combinations. I am just looking for uni-bi and tri grams. debug:{ rawquerystring:\best quality and best price\, querystring:\best quality and best price\, *parsedquery*:MultiPhraseQuery(text:\(best best_best quality best quality best quality _ best quality _ best) (quality quality _ quality _ best quality _ best price) (_ best _ best price _ best price_best) (best best_best price best price) price\), *parsedquery_toString*:text:\(best best_best quality best quality best quality _ best quality _ best) (quality quality _ quality _ best quality _ best price) (_ best _ best price _ best price_best) (best best_best price best price) price\, explain:{}, QParser:LuceneQParser, .. On Sun, Mar 23, 2014 at 11:31 AM, Jack Krupansky j...@basetechnology.comwrote: The query parser only presents the query terms one at a time to the analyzer, so your analyzer doesn't see both terms on one analysis call. If you enclose your query terms in quotes as a single phrase, you should see multiple terms being processed. q=best price -- Jack Krupansky -Original Message- From: Jyotirmoy Sundi Sent: Thursday, March 20, 2014 2:06 PM To: solr-user@lucene.apache.org Subject: Singles in solr for bigrams,trigrams in parsed_query Hi Folks, I am using singles to index bigrams/trigrams. The same is also used for query in the schema.xml file. But when I run the query in debug mode for a collections, I dont see the bigrams in the parsed_query . Any idea what I might be missing. solr/colection/select?q=best%20pricedebugQuery=on str name=parsedquery_toStringtext:best text:price/str I was hoping to see str name=parsedquery_toStringtext:best text:price text:best price/str My schema files looks like this: types fieldType name=string class=solr.StrField sortMissingLast=true omitNorms=true/ fieldType name=int class=solr.TrieIntField precisionStep=0 omitNorms=true positionIncrementGap=0/ fieldType name=text class=solr.TextField positionIncrementGap=100 analyzer type=index charFilter class=solr.HTMLStripCharFilterFactory/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=4 outputUnigrams=true / tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.LengthFilterFactory min=3 max=50 / filter class=solr.WordDelimiterFilterFactory generateWordParts=0 generateNumberParts=0 catenateWords=1 catenateNumbers=1 catenateAll=1 preserveOriginal=1 splitOnCaseChange=0 splitOnNumerics=0 stemEnglishPossessive=1/ filter class=solr.StopFilterFactory/ filter class=solr.SynonymFilterFactory synonyms=synonyms.txt ignoreCase=true expand=true/ filter class=solr.TrimFilterFactory / /analyzer analyzer type=query filter class=solr.LowerCaseFilterFactory/ filter class=solr.LengthFilterFactory min=3 max=50 / tokenizer class=solr.WhitespaceTokenizerFactory/ filter class=solr.StopFilterFactory/ filter class=solr.TrimFilterFactory / filter class=solr.SynonymFilterFactory synonyms=synonyms.txt ignoreCase=true expand=true/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=1 catenateNumbers=1 catenateAll=1 splitOnCaseChange=0 splitOnNumerics=0 stemEnglishPossessive=1/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=4 outputUnigrams=true / filter class=solr.CommonGramsFilterFactory words=stopwords.txt ignoreCase=true/ !--filter class=solr.CommonGramsFilterFactory words=stopwords.txt ignoreCase=true/ filter class=solr.ShingleFilterFactory minShingleSize=2 maxShingleSize=4 outputUnigrams=true /-- /analyzer /fieldType /types -- Best Regards, Jyotirmoy Sundi -- Best Regards, Jyotirmoy Sundi
Re: wrong results with wdf ngtf
What indexed text are you expecting the avaloq frage 1 query to match against? I just noticed that you have two distinct calls to WDF in your index analyzer. I think you're going to need to go back and clearly state all of the term requirements for both indexing and query. Show all the use cases, both index and query. You have too many balls in the air right now for anybody to be confident about what you're really trying to do. -- Jack Krupansky -Original Message- From: Andreas Owen Sent: Thursday, March 20, 2014 5:48 AM To: solr-user@lucene.apache.org Subject: wrong results with wdf ngtf Is there a way to tell ngramfilterfactory while indexing that number shall never be tokenized? then the query should be able to find numbers. Or do i have to change the ngram-min for numbers (not alpha) to 1, if that is possible? So to speak put the hole number as token and not all possible tokens. Solr analysis shows onnly WDF has no underscore in its tokens, the rest have it. can i tell the query to search numbers differently with NGTF, WT, LCF or whatever? I also tried filter class=solr.WordDelimiterFilterFactory types=at-under-alpha.txt/ @ = ALPHA _ = ALPHA I have gotten nearly everything to work. There are to queries where i dont get back what i want. avaloq frage 1 - only returns if i set minGramSize=1 while indexing yh_cug- query parser doesn't remove _ but the indexer does (WDF) so there is no match Is there a way to also query the hole term avaloq frage 1 without tokenizing it? Fieldtype: fieldType name=text_de class=solr.TextField positionIncrementGap=100 analyzer type=index tokenizer class=solr.StandardTokenizerFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.WordDelimiterFilterFactory types=at-under-alpha.txt/ filter class=solr.StopFilterFactory ignoreCase=true words=lang/stopwords_de.txt format=snowball enablePositionIncrements=true/ !-- remove common words -- filter class=solr.GermanNormalizationFilterFactory/ filter class=solr.SnowballPorterFilterFactory language=German/ !-- remove noun/adjective inflections like plural endings -- filter class=solr.NGramFilterFactory minGramSize=3 maxGramSize=15/ filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=1 catenateNumbers=1 catenateAll=0 splitOnCaseChange=1/ /analyzer analyzer type=query tokenizer class=solr.WhiteSpaceTokenizerFactory/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.WordDelimiterFilterFactory types=at-under-alpha.txt/ filter class=solr.StopFilterFactory ignoreCase=true words=lang/stopwords_de.txt format=snowball enablePositionIncrements=true/ !-- remove common words -- filter class=solr.GermanNormalizationFilterFactory/ filter class=solr.SnowballPorterFilterFactory language=German/ /analyzer /fieldType Solrconfig: queryParser name=synonym_edismax class=solr.SynonymExpandingExtendedDismaxQParserPlugin lst name=synonymAnalyzers lst name=myCoolAnalyzer lst name=tokenizer str name=classstandard/str /lst lst name=filter str name=classshingle/str str name=outputUnigramsIfNoShinglestrue/str str name=outputUnigramstrue/str str name=minShingleSize2/str str name=maxShingleSize4/str /lst lst name=filter str name=classsynonym/str str name=tokenizerFactorysolr.KeywordTokenizerFactory/str str name=synonymssynonyms.txt/str str name=expandtrue/str str name=ignoreCasetrue/str /lst /lst /lst /queryParser requestHandler name=/select2 class=solr.SearchHandler lst name=defaults str name=echoParamsexplicit/str int name=rows10/int str name=defTypesynonym_edismax/str str name=synonymstrue/str str name=qfplain_text^10 editorschoice^200 title^20 h_*^14 tags^10 thema^15 inhaltstyp^6 breadcrumb^6 doctype^10 contentmanager^5 links^5 last_modified^5 url^5 /str str name=bq(expiration:[NOW TO *] OR (*:* -expiration:*))^6/str str name=bfdiv(clicks,max(displays,1))^8/str !-- tested -- str name=dftext/str str name=fl*,path,score/str str name=wtjson/str str name=q.opAND/str !-- Highlighting defaults -- str name=hlon/str str
Re: [solr 4.7.0] analysis page: issue with HTMLStripCharFilterFactory
Stefan, no worries. The patch has fixed the issue. On Sun, Mar 16, 2014 at 10:39 PM, Stefan Matheis matheis.ste...@gmail.comwrote: Oh .. i'm sorry .. late to the party - didn't see the response from Doug .. so feel free to ignore that mail (: On Sunday, March 16, 2014 at 9:38 PM, Stefan Matheis wrote: Hey Dmitry We had a similar issue reported and already fixed: https://issues.apache.org/jira/browse/SOLR-5800 i'd suspect that this patch fixes your issue too? would like to hear back from you, if that's the case :) -Stefan On Saturday, March 15, 2014 at 6:58 PM, Dmitry Kan wrote: Hello, The following type does not get analyzed properly on the solr 4.7.0 analysis page: fieldType name=text_en_splitting class=solr.TextField positionIncrementGap=100 autoGeneratePhraseQueries=true analyzer type=index charFilter class=solr.HTMLStripCharFilterFactory/ !-- tokenizer class=solr.WhitespaceTokenizerFactory/ -- tokenizer class=solr.StandardTokenizerFactory / filter class=solr.StopFilterFactory ignoreCase=true words=lang/stopwords_en.txt / filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=1 catenateNumbers=1 catenateAll=0 splitOnCaseChange=1/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.KeywordMarkerFilterFactory protected=protwords.txt/ filter class=solr.PorterStemFilterFactory/ /analyzer analyzer type=query tokenizer class=solr.StandardTokenizerFactory / filter class=solr.SynonymFilterFactory synonyms=synonyms.txt ignoreCase=true expand=true/ filter class=solr.StopFilterFactory ignoreCase=true words=lang/stopwords_en.txt / filter class=solr.WordDelimiterFilterFactory generateWordParts=1 generateNumberParts=1 catenateWords=0 catenateNumbers=0 catenateAll=0 splitOnCaseChange=1/ filter class=solr.LowerCaseFilterFactory/ filter class=solr.KeywordMarkerFilterFactory protected=protwords.txt/ filter class=solr.PorterStemFilterFactory/ /analyzer /fieldType Example text: fox jumps Screenshot: http://pbrd.co/1lEVEIa This works fine in solr 4.6.1. -- Dmitry Blog: http://dmitrykan.blogspot.com Twitter: http://twitter.com/dmitrykan -- Dmitry Blog: http://dmitrykan.blogspot.com Twitter: http://twitter.com/dmitrykan
Re: SolrCloud from Stopping recovery for warnings to crash
Hi Lukas; Did you upgrade your architecture from an old version of Solr to Solr 4.7 or not? Thanks; Furkan KAMACI 23 Mar 2014 17:39 tarihinde Lukas Mikuckis lukasmikuc...@gmail.com yazdı: Solr version: 4.7 Architecture: 2 solrs (1 shard, leader + replica) 3 zookeepers Servers: * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores * zookeeper Solr data: * 21 collections * Many fields, small docs, docs count per collection from 1k to 500k About a week ago solr started crashing. It crashes every day, 3-4 times a day. Usually at nigh. I can't tell anything what could it be related to because at that time we haven't done any configuration changes. Load haven't changed too. Everything starts with Stopping recovery for .. warnings (every warnings is repeated several times): WARN org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for zkNodeName=core_node1core=** WARN org.apache.solr.cloud.ElectionContext; cancelElection did not find election node to remove WARN org.apache.solr.update.PeerSync; no frame of reference to tell if we've missed updates WARN - 2014-03-23 04:00:26.286; org.apache.solr.update.PeerSync; no frame of reference to tell if we've missed updates WARN - 2014-03-23 04:00:30.728; org.apache.solr.handler.SnapPuller; File _f9m_Lucene41_0.doc expected to be 6218278 while it is 7759879 WARN - 2014-03-23 04:00:54.126; org.apache.solr.update.UpdateLog$LogReplayer; Starting log replay tlog{file=/path/solr/collection1_shard1_replica2/data/tlog/tlog.0003272 refcount=2} active=true starting pos=356216606 Then again Stopping recovery for .. warnings: WARN org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for zkNodeName=core_node1core=** ERROR - 2014-03-23 05:19:29.566; org.apache.solr.common.SolrException; org.apache.solr.common.SolrException: No registered leader was found after waiting for 4000ms , collection: collection1 slice: shard1 ERROR - 2014-03-23 05:20:03.961; org.apache.solr.common.SolrException; org.apache.solr.common.SolrException: I was asked to wait on state down for IP:PORT_solr but I still do not see the requested state. I see state: active live:false After this serves mostly didn't recover.
Re: setting up solr on tomcat
What is the exception stack trace? The link looks good and works for Solr4.x From: Michael Sokolov msoko...@safaribooksonline.com To: solr-user@lucene.apache.org Sent: Sunday, March 23, 2014 7:56 AM Subject: Re: setting up solr on tomcat On 3/22/2014 2:16 AM, anupamk wrote: Hi, Is the solrTomcat wiki article valid for solr-4.7.0 ? http://wiki.apache.org/solr/SolrTomcat I am not able to deploy solr after following the instructions there. When I try to access the solr admin page I get a 404. I followed every step exactly as mentioned in the wiki, still no dice. Any ideas ? -- View this message in context: http://lucene.472066.n3.nabble.com/setting-up-solr-on-tomcat-tp4126177.html Sent from the Solr - User mailing list archive at Nabble.com. There was a note on that page saying: Solr4.3 http://wiki.apache.org/solr/Solr4.3requires completely different deployment. These instructions are*not*current and are for an indeterminate version of Solr. I haven't read the instructions in detail, but in my experience setting up a single standalone server goes like this: copy solr.war to the tomcat/webapps folder, logging jars (log4j, slf4j) and configuration (log4j.properties) to the tomcat/lib folder you can create your solr home directory directly in the tomcat folder -- if you do that, it should be found, or you can put it somewhere else and start the jvm with -Dsolr.solr.home=/wherever/you/put/solr that's pretty much it, I think. You will see the solr admin at http://localhost:8080/solr if you use all vanilla settings. -Mike
Re: SolrCloud from Stopping recovery for warnings to crash
Did you upgrade recently to Solr 4.7? 4.7 has a bad bug which can cause out of memory issues. Can you check your logs for out of memory errors? On Sun, Mar 23, 2014 at 9:07 PM, Lukas Mikuckis lukasmikuc...@gmail.com wrote: Solr version: 4.7 Architecture: 2 solrs (1 shard, leader + replica) 3 zookeepers Servers: * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores * zookeeper Solr data: * 21 collections * Many fields, small docs, docs count per collection from 1k to 500k About a week ago solr started crashing. It crashes every day, 3-4 times a day. Usually at nigh. I can't tell anything what could it be related to because at that time we haven't done any configuration changes. Load haven't changed too. Everything starts with Stopping recovery for .. warnings (every warnings is repeated several times): WARN org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for zkNodeName=core_node1core=** WARN org.apache.solr.cloud.ElectionContext; cancelElection did not find election node to remove WARN org.apache.solr.update.PeerSync; no frame of reference to tell if we've missed updates WARN - 2014-03-23 04:00:26.286; org.apache.solr.update.PeerSync; no frame of reference to tell if we've missed updates WARN - 2014-03-23 04:00:30.728; org.apache.solr.handler.SnapPuller; File _f9m_Lucene41_0.doc expected to be 6218278 while it is 7759879 WARN - 2014-03-23 04:00:54.126; org.apache.solr.update.UpdateLog$LogReplayer; Starting log replay tlog{file=/path/solr/collection1_shard1_replica2/data/tlog/tlog.0003272 refcount=2} active=true starting pos=356216606 Then again Stopping recovery for .. warnings: WARN org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for zkNodeName=core_node1core=** ERROR - 2014-03-23 05:19:29.566; org.apache.solr.common.SolrException; org.apache.solr.common.SolrException: No registered leader was found after waiting for 4000ms , collection: collection1 slice: shard1 ERROR - 2014-03-23 05:20:03.961; org.apache.solr.common.SolrException; org.apache.solr.common.SolrException: I was asked to wait on state down for IP:PORT_solr but I still do not see the requested state. I see state: active live:false After this serves mostly didn't recover. -- Regards, Shalin Shekhar Mangar.
Re: Solr4.7 No live SolrServers available to handle this request
Hi Greg, Where i can find the clusterstate.json.? i have ensemble zookeeper. Can you please tell me that where i can find it. Thanks. -- View this message in context: http://lucene.472066.n3.nabble.com/Solr4-7-No-live-SolrServers-available-to-handle-this-request-tp4125679p4126452.html Sent from the Solr - User mailing list archive at Nabble.com.