Re: How to stop backup once initiated

2014-03-23 Thread Shalin Shekhar Mangar
There is no way to stop it once initiated.

On Fri, Mar 21, 2014 at 8:09 PM, search engn dev
sachinyadav0...@gmail.com wrote:
 My index size 20 GB and I have issues solr backup command , now this backup
 is going on its taking too much time , so how can i stop backup command?



 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/How-to-stop-backup-once-initiated-tp4126020.html
 Sent from the Solr - User mailing list archive at Nabble.com.



-- 
Regards,
Shalin Shekhar Mangar.


Re: setting up solr on tomcat

2014-03-23 Thread Michael Sokolov

On 3/22/2014 2:16 AM, anupamk wrote:

Hi,

Is the solrTomcat wiki article valid for solr-4.7.0 ?
http://wiki.apache.org/solr/SolrTomcat


I am not able to deploy solr after following the instructions there.

When I try to access the solr admin page I get a 404.

I followed every step exactly as mentioned in the wiki, still no dice.

Any ideas ?



--
View this message in context: 
http://lucene.472066.n3.nabble.com/setting-up-solr-on-tomcat-tp4126177.html
Sent from the Solr - User mailing list archive at Nabble.com.

There was a note on that page saying:

Solr4.3 http://wiki.apache.org/solr/Solr4.3requires completely 
different deployment. These instructions are*not*current and are for an 
indeterminate version of Solr.


I haven't read the instructions in detail, but in my experience setting 
up a single standalone server goes like this:


copy solr.war to the tomcat/webapps folder,
logging jars (log4j, slf4j) and configuration (log4j.properties) to the 
tomcat/lib folder


you can create your solr home directory directly in the tomcat folder -- 
if you do that, it should be found, or you can put it somewhere else and 
start the jvm with -Dsolr.solr.home=/wherever/you/put/solr


that's pretty much it, I think.  You will see the solr admin at 
http://localhost:8080/solr if you use all vanilla settings.


-Mike


Re: understand debuginfo from query

2014-03-23 Thread Jack Krupansky
How are you getting the impression that is not worth anything? It appears to 
be worth more than the inhaltstyp field, but less than the title field - as 
you have specified by your field boosts.


-- Jack Krupansky

-Original Message- 
From: ao...@hispeed.ch

Sent: Friday, March 21, 2014 4:59 AM
To: solr-user@lucene.apache.org
Subject: Re: understand debuginfo from query

i found a good page to explain the debug output but it is still unclear for 
me. why is the field plain_text not worth anything? the query term was found 
3 times.


you can see it here: http://explain.solr.pl/explains/a90aze3o



 ao...@hispeed.ch schrieb:

i want the infos simplified so that the user can see why a doc was found

bellow is the output a a doc:

0.085597195 = (MATCH) sum of:
  0.083729245 = (MATCH) max of:
0.0019158133 = (MATCH) weight(plain_text:test^10.0 in 601) 
[DefaultSimilarity], result of:

  0.0019158133 = score(doc=601,freq=9.0 = termFreq=9.0
), product of:
0.022560213 = queryWeight, product of:
  10.0 = boost
  3.6232536 = idf(docFreq=81, maxDocs=1130)
  6.2265067E-4 = queryNorm
0.084920004 = fieldWeight in 601, product of:
  3.0 = tf(freq=9.0), with freq of:
9.0 = termFreq=9.0
  3.6232536 = idf(docFreq=81, maxDocs=1130)
  0.0078125 = fieldNorm(doc=601)
0.083729245 = (MATCH) weight(inhaltstyp:test^6.0 in 601) 
[DefaultSimilarity], result of:

  0.083729245 = score(doc=601,freq=1.0 = termFreq=1.0
), product of:
0.017686278 = queryWeight, product of:
  6.0 = boost
  4.734136 = idf(docFreq=26, maxDocs=1130)
  6.2265067E-4 = queryNorm
4.734136 = fieldWeight in 601, product of:
  1.0 = tf(freq=1.0), with freq of:
1.0 = termFreq=1.0
  4.734136 = idf(docFreq=26, maxDocs=1130)
  1.0 = fieldNorm(doc=601)
0.013458222 = (MATCH) weight(title:test^20.0 in 601) 
[DefaultSimilarity], result of:

  0.013458222 = score(doc=601,freq=1.0 = termFreq=1.0
), product of:
0.042281017 = queryWeight, product of:
  20.0 = boost
  3.395244 = idf(docFreq=102, maxDocs=1130)
  6.2265067E-4 = queryNorm
0.31830412 = fieldWeight in 601, product of:
  1.0 = tf(freq=1.0), with freq of:
1.0 = termFreq=1.0
  3.395244 = idf(docFreq=102, maxDocs=1130)
  0.09375 = fieldNorm(doc=601)
  0.001867952 = (MATCH) product of:
0.003735904 = (MATCH) sum of:
  0.003735904 = (MATCH) ConstantScore(expiration:[1395328539325 TO 
*]), product of:

1.0 = boost
0.003735904 = queryNorm
0.5 = coord(1/2)
  0.0 = (MATCH) 
FunctionQuery(div(int(clicks),max(int(displays),const(1, product of:

0.0 = div(int(clicks)=0,max(int(displays)=432,const(1)))
8.0 = boost
6.2265067E-4 = queryNorm


why is the sum 0.085597195? this would mean 0.083729245 + 0.001867952 and 
these are not included in the sum: 0.0019158133 + 0.013458222  + 
0.003735904


am i looking at the wrong total?
aren't these 2 cases the ones i have to sum up x = (MATCH) sum of or x = 
score( ?


i'm trying to extract the fields that where used for weighing the doc.



SolrCloud from Stopping recovery for warnings to crash

2014-03-23 Thread Lukas Mikuckis
Solr version: 4.7

Architecture:
2 solrs (1 shard, leader + replica)
3 zookeepers

Servers:
* zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores
* zookeeper + solr  (heap 4gb) - RAM 8gb, 2 cpu cores
* zookeeper

Solr data:
* 21 collections
* Many fields, small docs, docs count per collection from 1k to 500k

About a week ago solr started crashing. It crashes every day, 3-4 times a
day. Usually at nigh. I can't tell anything what could it be related to
because at that time we haven't done any configuration changes. Load
haven't changed too.


Everything starts with Stopping recovery for .. warnings (every warnings is
repeated several times):

WARN  org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for
zkNodeName=core_node1core=**

WARN  org.apache.solr.cloud.ElectionContext; cancelElection did not find
election node to remove

WARN  org.apache.solr.update.PeerSync; no frame of reference to tell if
we've missed updates

WARN  - 2014-03-23 04:00:26.286; org.apache.solr.update.PeerSync; no frame
of reference to tell if we've missed updates

WARN  - 2014-03-23 04:00:30.728; org.apache.solr.handler.SnapPuller; File
_f9m_Lucene41_0.doc expected to be 6218278 while it is 7759879

WARN  - 2014-03-23 04:00:54.126;
org.apache.solr.update.UpdateLog$LogReplayer; Starting log replay
tlog{file=/path/solr/collection1_shard1_replica2/data/tlog/tlog.0003272
refcount=2} active=true starting pos=356216606

Then again Stopping recovery for .. warnings:

WARN  org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for
zkNodeName=core_node1core=**

ERROR - 2014-03-23 05:19:29.566; org.apache.solr.common.SolrException;
org.apache.solr.common.SolrException: No registered leader was found after
waiting for 4000ms , collection: collection1 slice: shard1

ERROR - 2014-03-23 05:20:03.961; org.apache.solr.common.SolrException;
org.apache.solr.common.SolrException: I was asked to wait on state down for
IP:PORT_solr but I still do not see the requested state. I see state:
active live:false


After this serves mostly didn't recover.


Re: Shingles in solr for bigrams,trigrams in parsed_query

2014-03-23 Thread Jack Krupansky


The query parser only presents the query terms one at a time to the 
analyzer, so your analyzer doesn't see both terms on one analysis call.


If you enclose your query terms in quotes as a single phrase, you should see 
multiple terms being processed.


q=best price

-- Jack Krupansky

-Original Message- 
From: Jyotirmoy Sundi

Sent: Thursday, March 20, 2014 2:06 PM
To: solr-user@lucene.apache.org
Subject: Singles in solr for bigrams,trigrams in parsed_query

Hi Folks,
  I am using singles to index bigrams/trigrams. The same is also used
for query in the schema.xml file. But when I run the query in debug mode
for a collections, I dont see the bigrams in the parsed_query . Any idea
what I might be missing.
solr/colection/select?q=best%20pricedebugQuery=on

str name=parsedquery_toStringtext:best text:price/str
I was hoping to see
str name=parsedquery_toStringtext:best text:price text:best price/str

My schema files looks like this:
types
   fieldType name=string class=solr.StrField sortMissingLast=true
omitNorms=true/
   fieldType name=int class=solr.TrieIntField precisionStep=0
omitNorms=true positionIncrementGap=0/

   fieldType name=text class=solr.TextField
positionIncrementGap=100
 analyzer type=index
   charFilter class=solr.HTMLStripCharFilterFactory/
   filter class=solr.ShingleFilterFactory minShingleSize=2
maxShingleSize=4 outputUnigrams=true /
   tokenizer class=solr.WhitespaceTokenizerFactory/
   filter class=solr.LowerCaseFilterFactory/
   filter class=solr.LengthFilterFactory min=3 max=50 /
   filter class=solr.WordDelimiterFilterFactory
generateWordParts=0 generateNumberParts=0 catenateWords=1
catenateNumbers=1 catenateAll=1 preserveOriginal=1
splitOnCaseChange=0 splitOnNumerics=0 stemEnglishPossessive=1/
   filter class=solr.StopFilterFactory/
   filter class=solr.SynonymFilterFactory synonyms=synonyms.txt
ignoreCase=true expand=true/
   filter class=solr.TrimFilterFactory /
/analyzer

 analyzer type=query
   filter class=solr.LowerCaseFilterFactory/
   filter class=solr.LengthFilterFactory min=3 max=50 /
   tokenizer class=solr.WhitespaceTokenizerFactory/
   filter class=solr.StopFilterFactory/
   filter class=solr.TrimFilterFactory /
   filter class=solr.SynonymFilterFactory synonyms=synonyms.txt
ignoreCase=true expand=true/
   filter class=solr.WordDelimiterFilterFactory
generateWordParts=1 generateNumberParts=1 catenateWords=1
catenateNumbers=1 catenateAll=1 splitOnCaseChange=0
splitOnNumerics=0 stemEnglishPossessive=1/
   filter class=solr.ShingleFilterFactory minShingleSize=2
maxShingleSize=4 outputUnigrams=true /
   filter class=solr.CommonGramsFilterFactory words=stopwords.txt
ignoreCase=true/
   !--filter class=solr.CommonGramsFilterFactory
words=stopwords.txt ignoreCase=true/
   filter class=solr.ShingleFilterFactory minShingleSize=2
maxShingleSize=4 outputUnigrams=true /--
/analyzer
   /fieldType
/types



--
Best Regards,
Jyotirmoy Sundi 



Re: Multilingual indexing, search results, edismax and stopwords

2014-03-23 Thread Jack Krupansky
Setting the default query operator to AND is the preferred approach: 
q.op=AND.


That said, I'm not sure that counting ignored and empty terms towards the mm 
% makes sense. IOW, if a term transforms to nothing, either because it is a 
stop word or empty synonym replacement or pure punctuation, I don't think it 
should count as a term. I think this is worth a Jira.


-- Jack Krupansky

-Original Message- 
From: kastania44

Sent: Thursday, March 20, 2014 11:00 AM
To: solr-user@lucene.apache.org
Subject: Multilingual indexing, search results, edismax and stopwords

On our drupal multilingual system we use apache Solr 3.5.
The problem is well known on different blogs, sites I read.
The search results are not the one we want.
On our code in hook apachesolr_query_alter we override the defaultOperator:
$query-replaceParam('mm', '90%');
The requirement is, when I search for: biological analyses, I want to fetch
only the results which have both of the words.
When I search for: biological and chemical analyses, I want it to fetch only
the results which have biological , chemical, analyses. The and is not
indexed due to stopwords.

If I set mm to 100% and my query has stopwords it will not fetch any result.
If I set mm to 100$ and my query does not have stopwords it will fetch the
desired results.
If I set mm anything between 50%-99% it fetches not wanted results, as
results that contain only one of the searched keywords, or words like the
searched keywords, like analyse (even if I searched for analyses).

If I search using + before the words that are mandatory it works ok, but it
is not user friently, to ask from the user to type + before each word
exvcept from the stopwords.

Do I make any sense?

Below are some of our configuration details:

All the indexed fields are of type text_language,
e.g from our schema.xml
/field name=label type=text indexed=true stored=true
termVectors=true omitNorms=true/
field name=i18n_label_en type=text_en indexed=true stored=true
termVectors=true omitNorms=true/
field name=i18n_label_fr type=text_fr indexed=true stored=true
termVectors=true omitNorms=true//
All the text fieldtypes have the same configuration except from the
protected, words, dictionary parameters which are language specific.
e.g from our schema.xml
/fieldType name=text_en class=solr.TextField
positionIncrementGap=100
 analyzer type=index
   charFilter class=solr.MappingCharFilterFactory
mapping=mapping-ISOLatin1Accent_en.txt/
   tokenizer class=solr.WhitespaceTokenizerFactory/


   filter class=solr.StopFilterFactory ignoreCase=true
words=stopwords_en.txt enablePositionIncrements=true/
   filter class=solr.WordDelimiterFilterFactory
protected=protwords.txt generateWordParts=1 generateNumberParts=1
catenateWords=1 catenateNumbers=1 catenateAll=1 splitOnCaseChange=1
preserveOriginal=1 splitOnNumerics=1 stemEnglishPossessive=1/
   filter class=solr.LengthFilterFactory min=2 max=100/
   filter class=solr.LowerCaseFilterFactory/filter
class=solr.DictionaryCompoundWordTokenFilterFactory
dictionary=compoundwords_en.txt minWordSize=5 minSubwordSize=4
maxSubwordSize=15 onlyLongestMatch=true/
   filter class=solr.SnowballPorterFilterFactory language=English
protected=protwords_en.txt/
   filter class=solr.RemoveDuplicatesTokenFilterFactory/
 /analyzer
 analyzer type=query
   charFilter class=solr.MappingCharFilterFactory
mapping=mapping-ISOLatin1Accent_en.txt/
   tokenizer class=solr.WhitespaceTokenizerFactory/
   filter class=solr.SynonymFilterFactory synonyms=synonyms_en.txt
ignoreCase=true expand=true/
   filter class=solr.StopFilterFactory ignoreCase=true
words=stopwords_en.txt enablePositionIncrements=true/
   filter class=solr.WordDelimiterFilterFactory
protected=protwords.txt generateWordParts=1 generateNumberParts=1
catenateWords=1 catenateNumbers=1 catenateAll=1 splitOnCaseChange=1
preserveOriginal=1 splitOnNumerics=1 stemEnglishPossessive=1/
   filter class=solr.LengthFilterFactory min=2 max=100/
   filter class=solr.LowerCaseFilterFactory/filter
class=solr.DictionaryCompoundWordTokenFilterFactory
dictionary=compoundwords_en.txt minWordSize=5 minSubwordSize=4
maxSubwordSize=15 onlyLongestMatch=true/
   filter class=solr.SnowballPorterFilterFactory language=English
protected=protwords_en.txt/
   filter class=solr.RemoveDuplicatesTokenFilterFactory/
 /analyzer
   /fieldType/

solrQueryParser defaultOperator=AND/

solrconfig.xml

 /requestHandler name=pinkPony class=solr.SearchHandler
default=true
   lst name=defaults
 str name=defTypeedismax/str
 str name=echoParamsexplicit/str
 bool name=omitHeadertrue/bool
 float name=tie0.01/float

 int name=timeAllowed${solr.pinkPony.timeAllowed:-1}/int
 str name=q.alt*:*/str


 str name=spellcheckfalse/str

 str name=spellcheck.onlyMorePopulartrue/str
 str name=spellcheck.extendedResultsfalse/str

 str name=spellcheck.count1/str
   /lst
   arr 

Re: Shingles in solr for bigrams,trigrams in parsed_query

2014-03-23 Thread Jyotirmoy Sundi
Hi Jack,
  Thanks for your response, but if I try q=best quality and best
price, the parsedquery comes as following which is a lot of unwanted
combinations. I am just looking for uni-bi and tri grams.

debug:{
rawquerystring:\best quality and best price\,
querystring:\best quality and best price\,
*parsedquery*:MultiPhraseQuery(text:\(best best_best quality
best quality best quality _ best quality _ best) (quality quality _
quality _ best quality _ best price) (_ best _ best price _ best
price_best) (best best_best price best price) price\),
*parsedquery_toString*:text:\(best best_best quality best
quality best quality _ best quality _ best) (quality quality _ quality
_ best quality _ best price) (_ best _ best price _ best price_best)
(best best_best price best price) price\,
explain:{},
QParser:LuceneQParser,

..




On Sun, Mar 23, 2014 at 11:31 AM, Jack Krupansky j...@basetechnology.comwrote:


 The query parser only presents the query terms one at a time to the
 analyzer, so your analyzer doesn't see both terms on one analysis call.

 If you enclose your query terms in quotes as a single phrase, you should
 see multiple terms being processed.

 q=best price

 -- Jack Krupansky

 -Original Message- From: Jyotirmoy Sundi
 Sent: Thursday, March 20, 2014 2:06 PM
 To: solr-user@lucene.apache.org
 Subject: Singles in solr for bigrams,trigrams in parsed_query

 Hi Folks,
   I am using singles to index bigrams/trigrams. The same is also used
 for query in the schema.xml file. But when I run the query in debug mode
 for a collections, I dont see the bigrams in the parsed_query . Any idea
 what I might be missing.
 solr/colection/select?q=best%20pricedebugQuery=on

 str name=parsedquery_toStringtext:best text:price/str
 I was hoping to see
 str name=parsedquery_toStringtext:best text:price text:best
 price/str

 My schema files looks like this:
 types
fieldType name=string class=solr.StrField sortMissingLast=true
 omitNorms=true/
fieldType name=int class=solr.TrieIntField precisionStep=0
 omitNorms=true positionIncrementGap=0/

fieldType name=text class=solr.TextField
 positionIncrementGap=100
  analyzer type=index
charFilter class=solr.HTMLStripCharFilterFactory/
filter class=solr.ShingleFilterFactory minShingleSize=2
 maxShingleSize=4 outputUnigrams=true /
tokenizer class=solr.WhitespaceTokenizerFactory/
filter class=solr.LowerCaseFilterFactory/
filter class=solr.LengthFilterFactory min=3 max=50 /
filter class=solr.WordDelimiterFilterFactory
 generateWordParts=0 generateNumberParts=0 catenateWords=1
 catenateNumbers=1 catenateAll=1 preserveOriginal=1
 splitOnCaseChange=0 splitOnNumerics=0 stemEnglishPossessive=1/
filter class=solr.StopFilterFactory/
filter class=solr.SynonymFilterFactory synonyms=synonyms.txt
 ignoreCase=true expand=true/
filter class=solr.TrimFilterFactory /
 /analyzer

  analyzer type=query
filter class=solr.LowerCaseFilterFactory/
filter class=solr.LengthFilterFactory min=3 max=50 /
tokenizer class=solr.WhitespaceTokenizerFactory/
filter class=solr.StopFilterFactory/
filter class=solr.TrimFilterFactory /
filter class=solr.SynonymFilterFactory synonyms=synonyms.txt
 ignoreCase=true expand=true/
filter class=solr.WordDelimiterFilterFactory
 generateWordParts=1 generateNumberParts=1 catenateWords=1
 catenateNumbers=1 catenateAll=1 splitOnCaseChange=0
 splitOnNumerics=0 stemEnglishPossessive=1/
filter class=solr.ShingleFilterFactory minShingleSize=2
 maxShingleSize=4 outputUnigrams=true /
filter class=solr.CommonGramsFilterFactory words=stopwords.txt
 ignoreCase=true/
!--filter class=solr.CommonGramsFilterFactory
 words=stopwords.txt ignoreCase=true/
filter class=solr.ShingleFilterFactory minShingleSize=2
 maxShingleSize=4 outputUnigrams=true /--
 /analyzer
/fieldType
 /types



 --
 Best Regards,
 Jyotirmoy Sundi




-- 
Best Regards,
Jyotirmoy Sundi


Re: wrong results with wdf ngtf

2014-03-23 Thread Jack Krupansky
What indexed text are you expecting the avaloq frage 1 query to match 
against?


I just noticed that you have two distinct calls to WDF in your index 
analyzer.


I think you're going to need to go back and clearly state all of the term 
requirements for both indexing and query. Show all the use cases, both index 
and query. You have too many balls in the air right now for anybody to be 
confident about what you're really trying to do.


-- Jack Krupansky

-Original Message- 
From: Andreas Owen

Sent: Thursday, March 20, 2014 5:48 AM
To: solr-user@lucene.apache.org
Subject: wrong results with wdf  ngtf

Is there a way to tell ngramfilterfactory while indexing that number shall
never be tokenized? then the query should be able to find numbers.



Or do i have to change the ngram-min for numbers (not alpha) to 1, if that
is possible? So to speak put the hole number as token and not all possible
tokens.



Solr analysis shows onnly WDF has no underscore in its tokens, the rest have
it. can i tell the query to search numbers differently with NGTF, WT, LCF or
whatever?



I also tried filter class=solr.WordDelimiterFilterFactory
types=at-under-alpha.txt/

   @ = ALPHA

   _ = ALPHA



I have gotten nearly everything to work. There are to queries where i dont
get back what i want.



   avaloq frage 1   - only returns if i set
minGramSize=1 while indexing

   yh_cug- query parser doesn't
remove _ but the indexer does (WDF) so there is no match



Is there a way to also query the hole term avaloq frage 1 without
tokenizing it?



Fieldtype:



fieldType name=text_de class=solr.TextField positionIncrementGap=100

 analyzer type=index

  tokenizer
class=solr.StandardTokenizerFactory/

   filter
class=solr.LowerCaseFilterFactory/

  filter
class=solr.WordDelimiterFilterFactory types=at-under-alpha.txt/

  filter class=solr.StopFilterFactory
ignoreCase=true words=lang/stopwords_de.txt format=snowball
enablePositionIncrements=true/ !-- remove common words --

   filter
class=solr.GermanNormalizationFilterFactory/

  filter
class=solr.SnowballPorterFilterFactory language=German/ !-- remove
noun/adjective inflections like plural endings --


  filter class=solr.NGramFilterFactory
minGramSize=3 maxGramSize=15/

  filter
class=solr.WordDelimiterFilterFactory generateWordParts=1
generateNumberParts=1 catenateWords=1 catenateNumbers=1
catenateAll=0 splitOnCaseChange=1/

  /analyzer

  analyzer type=query

  tokenizer
class=solr.WhiteSpaceTokenizerFactory/

  filter
class=solr.LowerCaseFilterFactory/

  filter
class=solr.WordDelimiterFilterFactory types=at-under-alpha.txt/

  filter
class=solr.StopFilterFactory ignoreCase=true
words=lang/stopwords_de.txt format=snowball
enablePositionIncrements=true/ !-- remove common words --

  filter
class=solr.GermanNormalizationFilterFactory/

  filter
class=solr.SnowballPorterFilterFactory language=German/

 /analyzer

/fieldType





Solrconfig:




queryParser name=synonym_edismax



class=solr.SynonymExpandingExtendedDismaxQParserPlugin



  lst name=synonymAnalyzers



lst name=myCoolAnalyzer



  lst name=tokenizer



str name=classstandard/str



  /lst



  lst name=filter



str name=classshingle/str



str name=outputUnigramsIfNoShinglestrue/str



str name=outputUnigramstrue/str



str name=minShingleSize2/str



str name=maxShingleSize4/str



  /lst



  lst name=filter



str name=classsynonym/str



str name=tokenizerFactorysolr.KeywordTokenizerFactory/str



str name=synonymssynonyms.txt/str



str name=expandtrue/str



str name=ignoreCasetrue/str



  /lst



/lst



  /lst



/queryParser







requestHandler name=/select2 class=solr.SearchHandler



 lst name=defaults



   str name=echoParamsexplicit/str



   int name=rows10/int



   str name=defTypesynonym_edismax/str



   str name=synonymstrue/str



   str name=qfplain_text^10 editorschoice^200



title^20 h_*^14



tags^10 thema^15 inhaltstyp^6 breadcrumb^6 doctype^10



contentmanager^5 links^5



last_modified^5 url^5



   /str



   str name=bq(expiration:[NOW TO *] OR (*:*



-expiration:*))^6/str



   str name=bfdiv(clicks,max(displays,1))^8/str !-- tested --







   str name=dftext/str



   str name=fl*,path,score/str



   str name=wtjson/str



   str name=q.opAND/str







   !-- Highlighting defaults --



   str name=hlon/str



   str 

Re: [solr 4.7.0] analysis page: issue with HTMLStripCharFilterFactory

2014-03-23 Thread Dmitry Kan
Stefan,

no worries. The patch has fixed the issue.


On Sun, Mar 16, 2014 at 10:39 PM, Stefan Matheis
matheis.ste...@gmail.comwrote:

 Oh .. i'm sorry .. late to the party - didn't see the response from Doug
 .. so feel free to ignore that mail (:


 On Sunday, March 16, 2014 at 9:38 PM, Stefan Matheis wrote:

  Hey Dmitry
 
  We had a similar issue reported and already fixed:
 https://issues.apache.org/jira/browse/SOLR-5800
  i'd suspect that this patch fixes your issue too? would like to hear
 back from you, if that's the case :)
 
  -Stefan
 
  On Saturday, March 15, 2014 at 6:58 PM, Dmitry Kan wrote:
 
   Hello,
  
   The following type does not get analyzed properly on the solr 4.7.0
   analysis page:
  
   fieldType name=text_en_splitting class=solr.TextField
   positionIncrementGap=100 autoGeneratePhraseQueries=true
   analyzer type=index
   charFilter class=solr.HTMLStripCharFilterFactory/
   !-- tokenizer class=solr.WhitespaceTokenizerFactory/ --
   tokenizer class=solr.StandardTokenizerFactory /
   filter class=solr.StopFilterFactory
   ignoreCase=true
   words=lang/stopwords_en.txt
   /
   filter class=solr.WordDelimiterFilterFactory
   generateWordParts=1 generateNumberParts=1 catenateWords=1
   catenateNumbers=1 catenateAll=0 splitOnCaseChange=1/
   filter class=solr.LowerCaseFilterFactory/
   filter class=solr.KeywordMarkerFilterFactory
   protected=protwords.txt/
   filter class=solr.PorterStemFilterFactory/
   /analyzer
   analyzer type=query
   tokenizer class=solr.StandardTokenizerFactory /
   filter class=solr.SynonymFilterFactory synonyms=synonyms.txt
   ignoreCase=true expand=true/
   filter class=solr.StopFilterFactory
   ignoreCase=true
   words=lang/stopwords_en.txt
   /
   filter class=solr.WordDelimiterFilterFactory
   generateWordParts=1 generateNumberParts=1 catenateWords=0
   catenateNumbers=0 catenateAll=0 splitOnCaseChange=1/
   filter class=solr.LowerCaseFilterFactory/
   filter class=solr.KeywordMarkerFilterFactory
   protected=protwords.txt/
   filter class=solr.PorterStemFilterFactory/
   /analyzer
   /fieldType
  
   Example text:
   fox jumps
  
   Screenshot:
   http://pbrd.co/1lEVEIa
  
   This works fine in solr 4.6.1.
  
   --
   Dmitry
   Blog: http://dmitrykan.blogspot.com
   Twitter: http://twitter.com/dmitrykan
  
  
  
 
 




-- 
Dmitry
Blog: http://dmitrykan.blogspot.com
Twitter: http://twitter.com/dmitrykan


Re: SolrCloud from Stopping recovery for warnings to crash

2014-03-23 Thread Furkan KAMACI
Hi Lukas;

Did you upgrade your architecture from an old version of Solr to Solr 4.7
or not?

Thanks;
Furkan KAMACI
23 Mar 2014 17:39 tarihinde Lukas Mikuckis lukasmikuc...@gmail.com
yazdı:

 Solr version: 4.7

 Architecture:
 2 solrs (1 shard, leader + replica)
 3 zookeepers

 Servers:
 * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores
 * zookeeper + solr  (heap 4gb) - RAM 8gb, 2 cpu cores
 * zookeeper

 Solr data:
 * 21 collections
 * Many fields, small docs, docs count per collection from 1k to 500k

 About a week ago solr started crashing. It crashes every day, 3-4 times a
 day. Usually at nigh. I can't tell anything what could it be related to
 because at that time we haven't done any configuration changes. Load
 haven't changed too.


 Everything starts with Stopping recovery for .. warnings (every warnings is
 repeated several times):

 WARN  org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for
 zkNodeName=core_node1core=**

 WARN  org.apache.solr.cloud.ElectionContext; cancelElection did not find
 election node to remove

 WARN  org.apache.solr.update.PeerSync; no frame of reference to tell if
 we've missed updates

 WARN  - 2014-03-23 04:00:26.286; org.apache.solr.update.PeerSync; no frame
 of reference to tell if we've missed updates

 WARN  - 2014-03-23 04:00:30.728; org.apache.solr.handler.SnapPuller; File
 _f9m_Lucene41_0.doc expected to be 6218278 while it is 7759879

 WARN  - 2014-03-23 04:00:54.126;
 org.apache.solr.update.UpdateLog$LogReplayer; Starting log replay

 tlog{file=/path/solr/collection1_shard1_replica2/data/tlog/tlog.0003272
 refcount=2} active=true starting pos=356216606

 Then again Stopping recovery for .. warnings:

 WARN  org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for
 zkNodeName=core_node1core=**

 ERROR - 2014-03-23 05:19:29.566; org.apache.solr.common.SolrException;
 org.apache.solr.common.SolrException: No registered leader was found after
 waiting for 4000ms , collection: collection1 slice: shard1

 ERROR - 2014-03-23 05:20:03.961; org.apache.solr.common.SolrException;
 org.apache.solr.common.SolrException: I was asked to wait on state down for
 IP:PORT_solr but I still do not see the requested state. I see state:
 active live:false


 After this serves mostly didn't recover.



Re: setting up solr on tomcat

2014-03-23 Thread Pradeep Pujari
What is the exception stack trace? The link looks good and works for Solr4.x



 From: Michael Sokolov msoko...@safaribooksonline.com
To: solr-user@lucene.apache.org 
Sent: Sunday, March 23, 2014 7:56 AM
Subject: Re: setting up solr on tomcat
 

On 3/22/2014 2:16 AM, anupamk wrote:
 Hi,

 Is the solrTomcat wiki article valid for solr-4.7.0 ?
 http://wiki.apache.org/solr/SolrTomcat


 I am not able to deploy solr after following the instructions there.

 When I try to access the solr admin page I get a 404.

 I followed every step exactly as mentioned in the wiki, still no dice.

 Any ideas ?



 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/setting-up-solr-on-tomcat-tp4126177.html
 Sent from the Solr - User mailing list archive at Nabble.com.
There was a note on that page saying:

Solr4.3 http://wiki.apache.org/solr/Solr4.3requires completely 
different deployment. These instructions are*not*current and are for an 
indeterminate version of Solr.

I haven't read the instructions in detail, but in my experience setting 
up a single standalone server goes like this:

copy solr.war to the tomcat/webapps folder,
logging jars (log4j, slf4j) and configuration (log4j.properties) to the 
tomcat/lib folder

you can create your solr home directory directly in the tomcat folder -- 
if you do that, it should be found, or you can put it somewhere else and 
start the jvm with -Dsolr.solr.home=/wherever/you/put/solr

that's pretty much it, I think.  You will see the solr admin at 
http://localhost:8080/solr if you use all vanilla settings.

-Mike

Re: SolrCloud from Stopping recovery for warnings to crash

2014-03-23 Thread Shalin Shekhar Mangar
Did you upgrade recently to Solr 4.7? 4.7 has a bad bug which can
cause out of memory issues. Can you check your logs for out of memory
errors?

On Sun, Mar 23, 2014 at 9:07 PM, Lukas Mikuckis lukasmikuc...@gmail.com wrote:
 Solr version: 4.7

 Architecture:
 2 solrs (1 shard, leader + replica)
 3 zookeepers

 Servers:
 * zookeeper + solr (heap 4gb) - RAM 8gb, 2 cpu cores
 * zookeeper + solr  (heap 4gb) - RAM 8gb, 2 cpu cores
 * zookeeper

 Solr data:
 * 21 collections
 * Many fields, small docs, docs count per collection from 1k to 500k

 About a week ago solr started crashing. It crashes every day, 3-4 times a
 day. Usually at nigh. I can't tell anything what could it be related to
 because at that time we haven't done any configuration changes. Load
 haven't changed too.


 Everything starts with Stopping recovery for .. warnings (every warnings is
 repeated several times):

 WARN  org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for
 zkNodeName=core_node1core=**

 WARN  org.apache.solr.cloud.ElectionContext; cancelElection did not find
 election node to remove

 WARN  org.apache.solr.update.PeerSync; no frame of reference to tell if
 we've missed updates

 WARN  - 2014-03-23 04:00:26.286; org.apache.solr.update.PeerSync; no frame
 of reference to tell if we've missed updates

 WARN  - 2014-03-23 04:00:30.728; org.apache.solr.handler.SnapPuller; File
 _f9m_Lucene41_0.doc expected to be 6218278 while it is 7759879

 WARN  - 2014-03-23 04:00:54.126;
 org.apache.solr.update.UpdateLog$LogReplayer; Starting log replay
 tlog{file=/path/solr/collection1_shard1_replica2/data/tlog/tlog.0003272
 refcount=2} active=true starting pos=356216606

 Then again Stopping recovery for .. warnings:

 WARN  org.apache.solr.cloud.RecoveryStrategy; Stopping recovery for
 zkNodeName=core_node1core=**

 ERROR - 2014-03-23 05:19:29.566; org.apache.solr.common.SolrException;
 org.apache.solr.common.SolrException: No registered leader was found after
 waiting for 4000ms , collection: collection1 slice: shard1

 ERROR - 2014-03-23 05:20:03.961; org.apache.solr.common.SolrException;
 org.apache.solr.common.SolrException: I was asked to wait on state down for
 IP:PORT_solr but I still do not see the requested state. I see state:
 active live:false


 After this serves mostly didn't recover.



-- 
Regards,
Shalin Shekhar Mangar.


Re: Solr4.7 No live SolrServers available to handle this request

2014-03-23 Thread Sathya
Hi Greg,

Where i can find the clusterstate.json.? i have ensemble zookeeper. Can you
please tell me that where i can find it. 

Thanks.



--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr4-7-No-live-SolrServers-available-to-handle-this-request-tp4125679p4126452.html
Sent from the Solr - User mailing list archive at Nabble.com.