Re: How to improve this solr query?

2012-07-05 Thread Chamnap Chhorn
Hi Amit,

Thanks for your response.
1. It's just sometimes I see solr doesn't sort by score desc, so I made it
like that. I will have to check that again.
2. q1 and q2 are doing the search but just on different fields. String
fields means that it must match exactly, and solr need the q parameter to
be quoted. I did a nested query with the OR operator.

I'll check out the bf, pf, bq parameter more.

Thanks for the advise. :)

On Wed, Jul 4, 2012 at 2:28 PM, Amit Nithian anith...@gmail.com wrote:

 Couple questions:
 1) Why are you explicitly telling solr to sort by score desc,
 shouldn't it do that for you? Could this be a source of performance
 problems since sorting requires the loading of the field caches?
 2) Of the query parameters, q1 and q2, which one is actually doing
 text searching on your index? It looks like q1 is doing non-string
 related stuff, could this be better handled in either the bf or bq
 section of the edismax config? Looking at the sample though I don't
 understand how q1=apartment would hit non-string fields again (but see
 #3)
 3) Are the string fields literally of string type (i.e. no analysis
 on the field) or are you saying string loosely to mean text field.
 pf == phrase fields == given a multiple word query, will ensure that
 the specified phrase exists in the specified fields separated by some
 slop (hello my world may match hello world depending on this slop
 value). The qf means that given a multi term query, each term exists
 in the specified fields (name, description whatever text fields you
 want).

 Best
 Amit

 On Mon, Jul 2, 2012 at 9:35 AM, Chamnap Chhorn chamnapchh...@gmail.com
 wrote:
  Hi all,
 
  I'm using solr 3.5 with nested query on the 4 core cpu server + 17 Gb.
 The
  problem is that my query is so slow; the average response time is 12 secs
  against 13 millions documents.
 
  What I am doing is to send quoted string (q2) to string fields and
  non-quoted string (q1) to other fields and combine the result together.
 
 
 facet=truesort=score+descq2=*apartment*facet.mincount=1q1=*apartment*
 
 tie=0.1q.alt=*:*wt=jsonversion=2.2rows=20fl=uuidfacet.query=has_map:+truefacet.query=has_image:+truefacet.query=has_website:+truestart=0q=
  *
 
 _query_:+{!dismax+qf='.'+fq='..'+v=$q1}+OR+_query_:+{!dismax+qf='..'+fq='...'+v=$q2}
  *
 
 facet.field={!ex%3Ddt}sub_category_uuidsfacet.field={!ex%3Ddt}location_uuid
 
  I have done solr optimize already, but it's still slow. Any idea how to
  improve the speed? Am I done anything wrong?
 
  --
  Chhorn Chamnap
  http://chamnap.github.com/




-- 
Chhorn Chamnap
http://chamnap.github.com/


Search for abc AND *foo* return all docs for abc which do not have foo why?

2012-07-05 Thread Alok Bhandari
Hello,

If I Search for abc AND *foo* return all docs for abc which do not have foo
why? I suspect that if the * is present on both the side of a word then that
word is ignored. Is it the correct interpretation? I am using solr 3.6 and
field uses StandardTokenizer. 

--
View this message in context: 
http://lucene.472066.n3.nabble.com/Search-for-abc-AND-foo-return-all-docs-for-abc-which-do-not-have-foo-why-tp3993138.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: How to change tmp directory

2012-07-05 Thread Erik Fäßler
Ah - allright, that's it! Thank you!

Erik

Am 04.07.2012 um 17:59 schrieb Jack Krupansky:

 Solr is probably simply using Java's temp directory, which you can redefine 
 by setting the java.io.tmpdir system property on the java command line or 
 using a system-specific environment variable.
 
 -- Jack Krupansky
 
 -Original Message- From: Erik Fäßler
 Sent: Wednesday, July 04, 2012 3:56 AM
 To: solr-user@lucene.apache.org
 Subject: How to change tmp directory
 
 Hello all,
 
 I came about an odd issue today when I wanted to add ca. 7M documents to my 
 Solr index: I got a SolrServerException telling me No space left on device. 
 I had a look at the directory Solr (and its index) is installed in and there 
 is plenty space (~300GB).
 I then noticed a file named upload_457ee97b_1385125274b__8000_0005.tmp 
 had taken up all space of the machine's /tmp directory. The partition holding 
 the /tmp directory only has around 1GB of space and this file already took 
 nearly 800MB. I had a look at it and I realized that the file contained the 
 data I was adding to Solr in an XML format.
 
 Is there a possibility to change the temporary directory for this action?
 
 I use an IteratorSolrInputDocument with the HttpSolrServer's add(Iterator) 
 method for performance. So I can't just do commits from time to time.
 
 Best regards,
 
 Erik 



Re: Trunk error in Tomcat

2012-07-05 Thread Vadim Kisselmann
Hi Stefan,
ok, i would test the latest version from trunk with tomcat in next
days and open an new issue:)
regards
Vadim


2012/7/3 Stefan Matheis matheis.ste...@googlemail.com:
 On Tuesday, July 3, 2012 at 8:10 PM, Vadim Kisselmann wrote:
 sorry, i overlooked your latest comment with the new issue in SOLR-3238 ;)
 Should i open an new issue?


 NP Vadim, yes a new Issue would help .. all available Information too :)


Re: Get all matching terms of an OR query

2012-07-05 Thread Michael Jakl
Thank you!

On 4 July 2012 17:37, Jack Krupansky j...@basetechnology.com wrote:
 What exactly is it that is too slow?

I was comparing Queries with debugQuery enabled and disabled. The
difference was 60 seconds to 30 seconds for some (unusual) large
Queries (many Terms over a large set of documents chosen by filter
queries). After the caches are warm, the performance is of course far
better.

 It would be nice to have an optional search component or query parser option
 that returned the analyzed term for each query term.

Yes, I was thinking of reusing the analysis.jsp for that task, but
couldn't see an easy way to handle phrase queries and wasn't sure if
it performs better than the debugQuery approach.

 But as things stand, I would suggest that you do your own fuzzy match
 between the debugQuery terms and your source terms. That may not be 100%
 accurate, but probably would cover most/many cases.

Thanks, that's reassuring :)

Cheers,
Michael


Re: Search for abc AND *foo* return all docs for abc which do not have foo why?

2012-07-05 Thread Alok Bhandari
It is my mistake, the field which I was referring to was non existing so this
effect is shown. Sorry for the stupid question I have asked :-)

--
View this message in context: 
http://lucene.472066.n3.nabble.com/Search-for-abc-AND-foo-return-all-docs-for-abc-which-do-not-have-foo-why-tp3993138p3993147.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: alphanumeric interval

2012-07-05 Thread AlexR
Hi,

thanks a lot for your answer, and sorry for my late response.

It's my first time to write a solr plugin. I already have a plugin with
empty handleRequestBody() method and i'm able to call them.

I need the list of facetted field person (facet.field=person) in my method.
but i don't know how.

do you have a code snipped of your implementation?

thx
Alex
 

--
View this message in context: 
http://lucene.472066.n3.nabble.com/alphanumeric-interval-tp3990965p3993148.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Trunk error in Tomcat

2012-07-05 Thread Stefan Matheis
Great, thanks Vadim



On Thursday, July 5, 2012 at 9:34 AM, Vadim Kisselmann wrote:

 Hi Stefan,
 ok, i would test the latest version from trunk with tomcat in next
 days and open an new issue:)
 regards
 Vadim
 
 
 2012/7/3 Stefan Matheis matheis.ste...@googlemail.com 
 (mailto:matheis.ste...@googlemail.com):
  On Tuesday, July 3, 2012 at 8:10 PM, Vadim Kisselmann wrote:
   sorry, i overlooked your latest comment with the new issue in SOLR-3238 ;)
   Should i open an new issue?
  
  
  
  
  NP Vadim, yes a new Issue would help .. all available Information too :) 




Any ideas on Solr 4.0 Release.

2012-07-05 Thread Sohail Aboobaker
Hi,

Congratulations on Alpha release. I am wondering is there a ball park on
final release for 4.0? Is it expected in August or Sep time frame or is it
further away? We badly need some features included in this release. These
are around grouped facet counts. We have limited use for Solr in our
current release. In next release, we will add more features (full text
searching, location based searches etc.). I am wondering if the facet and
group counts side of things is stable in Alpha or not? I have tested with
the nightly builds before and it works fine for our scenarios.

Thanks.

Regards,
Sohail


Frequency of Unique Id displayed more than 1

2012-07-05 Thread Sohail Aboobaker
Hi,

We have defined a unique key as schemaid. We add documents using
server.addBean(obj) method. We are using the same method for updates as
well. When browsing the schema, we see that some of the schemaid values
have frequency of more than 1. Since, schemaid column is defined as unique
key, we are expecting when addBean, it will automatically replace the
existing entry in index.

Are we supposed to use a different method for update as opposed to add?

Regards,
Sohail


Re: Frequency of Unique Id displayed more than 1

2012-07-05 Thread Savvas Andreas Moysidis
Hello,

Make sure your unique id has a type which always yields one token
after tokenisation is applied (e.g. either string or a type which
only defines the KeywordTokenizer in its chain)

Regards,
Savvas

On 5 July 2012 11:02, Sohail Aboobaker sabooba...@gmail.com wrote:
 Hi,

 We have defined a unique key as schemaid. We add documents using
 server.addBean(obj) method. We are using the same method for updates as
 well. When browsing the schema, we see that some of the schemaid values
 have frequency of more than 1. Since, schemaid column is defined as unique
 key, we are expecting when addBean, it will automatically replace the
 existing entry in index.

 Are we supposed to use a different method for update as opposed to add?

 Regards,
 Sohail


Re: Frequency of Unique Id displayed more than 1

2012-07-05 Thread Sohail Aboobaker
We have defined the schemaid as String. It has concatenated value of the
product id and language. It takes the form of ID-EN. For example:
'123012-EN', '124020-EN', '12392-FR'.

Sohail


Re: Frequency of Unique Id displayed more than 1

2012-07-05 Thread Sohail Aboobaker
Another observation is that when we query an individual schemaid, it
returns only one row using the search interface. Why would frequency be
more than 1?


Re: Frequency of Unique Id displayed more than 1

2012-07-05 Thread Savvas Andreas Moysidis
can you post the schema you are applying pls?

On 5 July 2012 11:28, Sohail Aboobaker sabooba...@gmail.com wrote:
 Another observation is that when we query an individual schemaid, it
 returns only one row using the search interface. Why would frequency be
 more than 1?


Re: Use of Solr as primary store for search engine

2012-07-05 Thread Sohail Aboobaker
In many e-commerce sites, most of data that we display (except images)
especially in grids and lists is minimal. We were inclined to use Solr as
data store for only displaying the information in grids. We stopped only
due to non-availability of joins in Solr3.5. Since, our data (like any
other relational store) is split in multiple tables, we needed to
de-normalize to use solr as a store. We decided against it because that
would mean potentially heavy updates to indexes whenever related data is
updated. With Solr 4.0, we might have decided differently and implement the
grids using joins within solr.

We are too new to Solr to have any insights into it.

Regards,
Sohail


Synonyms and Regions Taxonomy

2012-07-05 Thread Stephen Lacy
When a user types in South America they want to be able to see documents
containing Brazil, Chile etc.
No I have already thrown together a list of countries and continents
however I'm a little more ambitious,
I would like to get a lot more regions such as american states as well or
Former members of the USSR...
Are there ready made synonym files or taxonomies in a different format.
Are synonyms the best way of achieving this? Perhaps there is a better way?
Any pitfalls or advice on this subject from someone who has done this
before would be appreciated.
Thanks

Stephen


solr-4.0 and high cpu usage

2012-07-05 Thread Anatoli Matuskova
Hello,
I'm testing solr-4.0-alpha compared to 1.4. My index is optimized to one
segment. I've seen a decrease in memory usage but a very high increase in
CPU. This high cpu usage ends up giving me slower response times with 4.0
than 1.4
The server I'm using: Linux version 2.6.30-2-amd64 16 2.26GHz Intel Pentium
Xeon Processors, 8GB RAM
I have a jmeter sending queries using between 1 and 4 threads.
The queries doesn't use faceting, neighter filters. Simple search to 5 text
fields using dismax
The index has 1M docs and is 1.2Gb size
I've checked memory with jvmconsole and it's not GC fault.
The index is built using the TieredMergePolicy for 4.0 and
LogByteSizeMergePolicy for 1.4 but both were optimized to 1 segment (so in
that case the merge policy shouldn't make any difference, am I correct?).
I'm not indexing any docs while doing the tests
Average response time came from 0.1sec to 0.3 sec

Here is a graph of the cpu increase:
Any advice or something I should take into account? With the same resources
solr-4.0 is being 3 times slower than 1.4 and I don't know what I'm doing
wrong




http://lucene.472066.n3.nabble.com/file/n3993187/ganglia-solr.png 

--
View this message in context: 
http://lucene.472066.n3.nabble.com/solr-4-0-and-high-cpu-usage-tp3993187.html
Sent from the Solr - User mailing list archive at Nabble.com.


Problems with elevation component configuration

2012-07-05 Thread Igor Salma
Hi to all,

we are using solr in combination with nutch and there are multiple cores
defined under solr. From some reason we can't configure elevation request
handler. We followed the instruction on
http://wiki.apache.org/solr/QueryElevationComponent,
only changed node for queryFieldType to work with text type fields:
searchComponent name=elevator class=solr.QueryElevationComponent 
str name=queryFieldTypetext/str
str name=config-fileelevate.xml/str
  /searchComponent

  requestHandler name=/elevate class=solr.SearchHandler startup=lazy
lst name=defaults
  str name=echoParamsexplicit/str
/lst
arr name=last-components
  strelevator/str
/arr
  /requestHandler

and added the node in elevate.xml :

query text=brain
doc id=
http://10.237.119.179:28080/ncal/mdo/presentation/conditions/conditionpage.jsp?condition=Condition_Epilepsy.xml;
/
 /query

(id field is the type of string and, as you can see, we are storing ulrs
as identifiers)

And when restart solr and try
http://localhost:8080/solr/elevate?q=indexingabstract:braindebugQuery=trueenableElevation=truenothing
changes - the mentioned document is not on the top of the resut
list.

Can someone please help?

Thanks in advance,
Igor


Re: data Model/UML diagrams/Class diagrams.

2012-07-05 Thread prasad deshpande
Is there any who can help me?

On Thu, Jul 5, 2012 at 10:10 AM, prasad deshpande 
prasad.deshpand...@gmail.com wrote:

 I would like to understand how solr/lucene/Tika frameworks designed. Where
 would I get the class diagram/UML diagrams to better understand the
 solr/lucene/Tika.
 I have the source code for all of them.


 Thanks,
 Prasad



Re: How negative queries are handled with edismax parser?

2012-07-05 Thread Erick Erickson
Have you looked at the results with debugQuery=on? I admit it takes a while
to get the hang of reading the debug output, but it will probably give
you a good
idea of what's going on

Best
Erick

On Tue, Jul 3, 2012 at 7:46 AM, Alok Bhandari
alokomprakashbhand...@gmail.com wrote:
 Hello,

 I am using and edismax parser, when I query for

 *- *  it returns 0 results but if I use lucene parser then I get all docs
 from the index. So want to know how exactly edismax handles this query and
 is it related to negative query? Please let me know.I am using solr 3.6

 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/How-negative-queries-are-handled-with-edismax-parser-tp3992724.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: Boosting the whole documents

2012-07-05 Thread Erick Erickson
Can we see the query and the output with debugQuery=on? Look
especially for ConstantScoreQuery in the debug output, if you
do certain searches then all the boosting stuff is ignored, e.g.
*:*.

Best
Erick

On Wed, Jul 4, 2012 at 9:39 AM, Danilak Michal michal.dani...@prezi.com wrote:
 Hi,

 I have the following problem.
 I would like to give a boost to the whole documents as I index them. I am
 sending to solr xml in the form:

 adddoc boost=2.0/doc/add

 But it does't seem to alter the search scores in any way. I would expect
 that to multiply the final search score by two, am I correct?
 Probably I would need to alter schema.xml, but I found only information on
 how to do that for specific fields (just put omitNorms=false into the field
 tag). But what should I do, if I want to boost the whole document?

 Note: by boosting a whole document I mean, that if document A has search
 score 10.0 and document B has search score 15.0 and I give document A the
 boost 2.0, when I index it, I would expect its search score to be 20.0.

 Thanks in advance!

 Michal Danilak


Re: Solr facet multiple constraint

2012-07-05 Thread Erick Erickson
Well, to start with this query is totally messed up.

facet=true,sort=publishingdate desc,facet.mincount=1,q=service:1 AND
publicationstatus:LIVE,facet.field={!ex=dt}user,wt=javabin,fq={!tag=dt}user:10,version=2

You've put in commas where you should have ampersands to separate
parameters=value pairs.

e.g.
facet=truesort=publishingdate descfacet.mincount=1q=service:1 AND
publicationstatus:LIVEfacet.field={!ex=dt}user,wt=javabin,fq={!tag=dt}user:10,version=2


Best
Erick

On Thu, Jul 5, 2012 at 1:17 AM, davidbougearel
david.bougea...@smile-benelux.com wrote:
 Please someone can help me,

 we are a team waiting for a fix.
 We try several ways to implement it without success.

 Thanks for reading anyway, David.

 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Solr-facet-multiple-constraint-tp3992974p3993119.html
 Sent from the Solr - User mailing list archive at Nabble.com.


RE: Any ideas on Solr 4.0 Release.

2012-07-05 Thread Steven A Rowe
Hi Sohail,

Some of your questions are answered here: 
http://wiki.apache.org/solr/Solr4.0. 

See Chris Hostetter's blog post for more info, particularly on questions around 
stability: 
http://www.lucidimagination.com/blog/2012/07/03/4-0-alpha-whats-in-a-name/.

Steve 

-Original Message-
From: Sohail Aboobaker [mailto:sabooba...@gmail.com] 
Sent: Thursday, July 05, 2012 5:22 AM
To: solr-user@lucene.apache.org
Subject: Any ideas on Solr 4.0 Release.

Hi,

Congratulations on Alpha release. I am wondering is there a ball park on final 
release for 4.0? Is it expected in August or Sep time frame or is it further 
away? We badly need some features included in this release. These are around 
grouped facet counts. We have limited use for Solr in our current release. In 
next release, we will add more features (full text searching, location based 
searches etc.). I am wondering if the facet and group counts side of things is 
stable in Alpha or not? I have tested with the nightly builds before and it 
works fine for our scenarios.

Thanks.

Regards,
Sohail


Re: data Model/UML diagrams/Class diagrams.

2012-07-05 Thread Anuj Kumar
Hi Prasad,

If you have the source code, why not reverse engineer the UML diagrams-
http://stackoverflow.com/questions/51786/recommended-eclipse-plugins-to-generate-uml-from-java-code

Regards,
Anuj

On Thu, Jul 5, 2012 at 6:42 PM, prasad deshpande 
prasad.deshpand...@gmail.com wrote:

 Is there any who can help me?

 On Thu, Jul 5, 2012 at 10:10 AM, prasad deshpande 
 prasad.deshpand...@gmail.com wrote:

  I would like to understand how solr/lucene/Tika frameworks designed.
 Where
  would I get the class diagram/UML diagrams to better understand the
  solr/lucene/Tika.
  I have the source code for all of them.
 
 
  Thanks,
  Prasad
 



Re: Frequency of Unique Id displayed more than 1

2012-07-05 Thread Erick Erickson
solr updates are really a delete followed by a re-index. The old terms
are left in the index, but the associated document is marked as deleted.
Schema browser, for instance, will happily report frequencies  1 for
uniqueKeys when a document has been updated.

You can ignore this if you query on the schemaId in question and get
only back one document, it's expected behavior. Also, if this is indeed
what you're seeing, doing  an optimize (forceMerge in the new parlance)
should take all the frequencies back to 1.

Best
Erick

On Thu, Jul 5, 2012 at 9:09 AM, Savvas Andreas Moysidis
savvas.andreas.moysi...@gmail.com wrote:
 hmm, based on the schema I don't see how you would be able to commit
 the same schemaid twice?
 maybe you want to investigate how you post a document to solr (do you
 do a commit after the post etc) or the merge strategy that is being
 applied.

 Just to exclude any possibilities, is is possible that at some point
 the schemaid field wasn't defined as string and then changed to
 string without re-indexing?

 On 5 July 2012 12:09, Sohail Aboobaker sabooba...@gmail.com wrote:
 fields
  field name=schemaid type=string indexed=true stored=true
 required=true/
  field name=itemmasterid type=string indexed=true stored=true
 required=true/
  field name=itemtitle type=string indexed=true stored=true
 required=true/
  field name=manufacturerId type=string indexed=true stored=true
 required=true/
 field name=language type=string indexed=true stored=true
 required=true/
 field name=make type=string indexed=true stored=true
 required=true/
 field name=model type=string indexed=true stored=true
 required=true/
 field name=msrp type=double indexed=true stored=true
 required=true/
 field name=level1Categories type=string indexed=true stored=true
 required=true multiValued=true/field name=level2Categories
 type=string indexed=true stored=true required=false
 multiValued=true/field name=level3Categories type=string
 indexed=true stored=true required=false multiValued=true/field
 name=level4Categories type=string indexed=true stored=true
 required=false multiValued=true/
 /fields
 uniqueKeyschemaid/uniqueKey

 Above is the main schema. Let me know if you need more information.


Re: solr-4.0 and high cpu usage [SOLVED]

2012-07-05 Thread Anatoli Matuskova
Found why!
On Solr 1.4 dismax param mm defaults to 1 if not specified, which is
equivalent to AND. On Solr 4.0 if mm is not specified, the default operator
is used, which defaults to OR. That made return much more results for each
query I was running, increasing the response time and the CPU usage.


--
View this message in context: 
http://lucene.472066.n3.nabble.com/solr-4-0-and-high-cpu-usage-tp3993187p3993275.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Problems with elevation component configuration

2012-07-05 Thread Chris Warner
Hi, Igor,

I also set forceElevation to true for my elevate results.

 Cheers,
Chris



- Original Message -
From: Igor Salma igor.sa...@gmail.com
To: solr-user@lucene.apache.org
Cc: 
Sent: Thursday, July 5, 2012 4:47 AM
Subject: Problems with elevation component configuration

Hi to all,

we are using solr in combination with nutch and there are multiple cores
defined under solr. From some reason we can't configure elevation request
handler. We followed the instruction on
http://wiki.apache.org/solr/QueryElevationComponent,
only changed node for queryFieldType to work with text type fields:
searchComponent name=elevator class=solr.QueryElevationComponent 
    str name=queryFieldTypetext/str
    str name=config-fileelevate.xml/str
  /searchComponent

  requestHandler name=/elevate class=solr.SearchHandler startup=lazy
    lst name=defaults
      str name=echoParamsexplicit/str
    /lst
    arr name=last-components
      strelevator/str
    /arr
  /requestHandler

and added the node in elevate.xml :

query text=brain
    doc id=
http://10.237.119.179:28080/ncal/mdo/presentation/conditions/conditionpage.jsp?condition=Condition_Epilepsy.xml;
/
/query

(id field is the type of string and, as you can see, we are storing ulrs
as identifiers)

And when restart solr and try
http://localhost:8080/solr/elevate?q=indexingabstract:braindebugQuery=trueenableElevation=truenothing
changes - the mentioned document is not on the top of the resut
list.

Can someone please help?

Thanks in advance,
Igor



Re: Frequency of Unique Id displayed more than 1

2012-07-05 Thread Sohail Aboobaker
Thanks Eric,

This is indeed what we are seeing. I hope we can just ignore the
frequencies. Does it in any way effect facet counts for such records?

Sohail


Re: alphanumeric interval

2012-07-05 Thread Cat Bieber
I did not use facets in my implementation, so I don't have any 
facet-specific code snippet that would be helpful to you. However, if 
your handler extends SearchHandler and calls super.handleRequestBody() 
it should be running the facet component code. You have access to the 
SolrQueryResponse built by it, and may be able to get the data out of 
that object. You'll need to look at the javadoc for NamedList, and I 
found it helpful to dump the list in debug statements so I could examine 
its structure and contents. I suspect you need something like 
rsp.getValues().get(facet_counts) to get the facet data, but haven't 
tested it.

-Cat Bieber

On 07/05/2012 04:32 AM, AlexR wrote:

Hi,

thanks a lot for your answer, and sorry for my late response.

It's my first time to write a solr plugin. I already have a plugin with
empty handleRequestBody() method and i'm able to call them.

I need the list of facetted field person (facet.field=person) in my method.
but i don't know how.

do you have a code snipped of your implementation?

thx
Alex


--
View this message in context: 
http://lucene.472066.n3.nabble.com/alphanumeric-interval-tp3990965p3993148.html
Sent from the Solr - User mailing list archive at Nabble.com.
   


Update JSON format post 3.1?

2012-07-05 Thread Klostermeyer, Michael
Is there any official documentation on the JSON format Solr 3.5 expects when 
adding/updating documents via /update/json?  Most of the official documentation 
is 3.1, and I am of the understanding this changed in v3.2 
(https://issues.apache.org/jira/browse/SOLR-2496).  I believe I have the 
correct format, but I am getting an odd error:

The request sent by the client was syntactically incorrect (Expected: 
OBJECT_START but got ARRAY_START

My JSON format is as follows (simplified):
{add:
{doc:
[
{ID:987654321,Name:Steve Smith,ChildIDs:[3841]}
]
}
}

The idea is that I want to be able to send multiple documents within the same 
request, although in this example I am demonstrating only a single document. 
ChildIDs is defined as a multivalued field.

Thanks.

Mike Klostermeyer


Re: Frequency of Unique Id displayed more than 1

2012-07-05 Thread Erick Erickson
No, it doesn't affect facet counts, grouping, or any of that stuff. After
all, facets and grouping are only calculated for documents that
satisfy a query, and deleted documents are, by definition, excluded
from satisfying a query.

There are some subtle issues in scoring that can be affected, but you'll
rarely care about that. And the extra information is purged over time. When
segment merges happen, the data associated with deleted documents
is removed from the segments being merge (which, in effect, is what you
force with an optimize)..

BTW, optimizing is rarely required, usually people only optimize when an index
is pretty static, that is more toward the write-once end of the spectrum.

Best
Erick

On Thu, Jul 5, 2012 at 1:44 PM, Sohail Aboobaker sabooba...@gmail.com wrote:
 Thanks Eric,

 This is indeed what we are seeing. I hope we can just ignore the
 frequencies. Does it in any way effect facet counts for such records?

 Sohail


Re: solr-4.0 and high cpu usage

2012-07-05 Thread Erick Erickson
This is rather surprising. I know some defaults have changed between
1.4 and 4.0,
so the first thing I'd do is try some of your sample queries with
debugQuery=on and
compare the debug info to see if you're executing more complex queries without
meaning to.

Also, dismax is deprecated in favor of edismax, so I'm not quite sure
you're comparing
apples/apples here.

But it's worth pursuing. And I'm assuming you're measuring QTime
rather than just
measuring response time, although you did say jmeter, so I'd guess
you're getting
response times, not QTime.

Best
Erick

On Thu, Jul 5, 2012 at 7:18 AM, Anatoli Matuskova
anatoli.matusk...@gmail.com wrote:
 Hello,
 I'm testing solr-4.0-alpha compared to 1.4. My index is optimized to one
 segment. I've seen a decrease in memory usage but a very high increase in
 CPU. This high cpu usage ends up giving me slower response times with 4.0
 than 1.4
 The server I'm using: Linux version 2.6.30-2-amd64 16 2.26GHz Intel Pentium
 Xeon Processors, 8GB RAM
 I have a jmeter sending queries using between 1 and 4 threads.
 The queries doesn't use faceting, neighter filters. Simple search to 5 text
 fields using dismax
 The index has 1M docs and is 1.2Gb size
 I've checked memory with jvmconsole and it's not GC fault.
 The index is built using the TieredMergePolicy for 4.0 and
 LogByteSizeMergePolicy for 1.4 but both were optimized to 1 segment (so in
 that case the merge policy shouldn't make any difference, am I correct?).
 I'm not indexing any docs while doing the tests
 Average response time came from 0.1sec to 0.3 sec

 Here is a graph of the cpu increase:
 Any advice or something I should take into account? With the same resources
 solr-4.0 is being 3 times slower than 1.4 and I don't know what I'm doing
 wrong




 http://lucene.472066.n3.nabble.com/file/n3993187/ganglia-solr.png

 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/solr-4-0-and-high-cpu-usage-tp3993187.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: what is the sequence of execution of solr query , is it right to left?

2012-07-05 Thread Erick Erickson
This question can't be answered correctly as asked. The query parser does not
implement boolean logic directly, so any answer is pretty much wrong. Here's
an excellent writeup of what the query parser actually does:

http://www.lucidimagination.com/blog/2011/12/28/why-not-and-or-and-not/

As pointed out in the blog if you were going to try to get
boolean-like behavior,
you must be careful to parenthesize properly

Best
Erick

On Thu, Jul 5, 2012 at 6:51 AM, Alok Bhandari
alokomprakashbhand...@gmail.com wrote:
 Hello,

 if I search solr with the criteria A AND B OR C , then what is the order of
 execution of boolean operators?

 I guess it is

 1)Get result of B OR C
 2) Get Result of A AND (result of step 1)

 Is it correct?. I am using solr 3.6.

 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/what-is-the-sequence-of-execution-of-solr-query-is-it-right-to-left-tp3993182.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: Synonyms and Regions Taxonomy

2012-07-05 Thread Tri Cao
I don't think there's a synonym file for this use case. I am not even sure if
synonym is the right way to handle it.

I think the better way to improve recall is to mark up your documents with
a hidden field of is the geographic relations. For example, before indexing,
you can add a field to all documents containing South America, something
like: South America is a subcontinent, that is consisted of the countries 
Brazil,
Chile, Argentina, …

This data can come from various sources, such as wikipedia, wordnet, etc.


On Jul 5, 2012, at 4:12 AM, Stephen Lacy wrote:

 When a user types in South America they want to be able to see documents
 containing Brazil, Chile etc.
 No I have already thrown together a list of countries and continents
 however I'm a little more ambitious,
 I would like to get a lot more regions such as american states as well or
 Former members of the USSR...
 Are there ready made synonym files or taxonomies in a different format.
 Are synonyms the best way of achieving this? Perhaps there is a better way?
 Any pitfalls or advice on this subject from someone who has done this
 before would be appreciated.
 Thanks
 
 Stephen



Re: Filtering a query by range returning unexpected results

2012-07-05 Thread Andrew Meredith
Thanks! That worked. I re-built the index with my prices field being of
the tfloat type, and I am now able to perform range queries. I appreciate
your help, Erick.

On Tue, Jul 3, 2012 at 9:35 AM, Erick Erickson erickerick...@gmail.comwrote:

 OK, this appears to be something with the currency type. It works fine
 for
 regular float fields. I can't get the multiValued currency types to work
 with
 range queries. Don't quite know what I was doing when I thought they
 _did_ work.

 One work-around I think, if you are using a single currency USD might be
 to copy your price to a simple float field and do your range queries on
 that.

 I'm not at all sure that the currency type was ever intended to support
 multiValued=true. I don't know enough about the internals to know if
 it's even a good idea to try, but the current behavior could be improved
 upon.

 But it seems to me that one of two things should happen:
 1 the startup should barf if a currency type is multiValued (fail early)
 or
 2 currency should work when multiValued.

 Unfortunately, JIRA is down so I can't look to see if this is already a
 known
 issue or enter a JIRA if it isn't. I'll try to look later if it all
 comes back up.

 Best
 Erick

 On Mon, Jul 2, 2012 at 1:53 PM, Andrew Meredith andymered...@gmail.com
 wrote:
  Yep, that 15.00.00 was a typo.
 
  Here are the relevant portions of my schema.xml:
 
  types
  !-- CUT --
  fieldType name=currency class=solr.CurrencyField
 precisionStep=8
  defaultCurrency=USD currencyConfig=currency.xml /
  !-- CUT --
  /types
 
  fields
  !-- CUT --
  field name=prices type=currency indexed=true stored=true
  multiValued=true /
  !-- CUT --
  /fields
 
  And here is the output of a sample query with debugQuery=on appended:
 
  lst name=debug
  str name=rawquerystringFurtick/str
  str name=querystringFurtick/str
  str name=parsedquery
  +DisjunctionMaxQuery((subtitle:furtick | frontlist_flapcopy:furtick^0.5 |
  frontlist_ean:furtick^6.0 | author:furtick^3.0 | series:furtick^1.5 |
  title:furtick^2.0)) ()
  /str
  str name=parsedquery_toString
  +(subtitle:furtick | frontlist_flapcopy:furtick^0.5 |
  frontlist_ean:furtick^6.0 | author:furtick^3.0 | series:furtick^1.5 |
  title:furtick^2.0) ()
  /str
  lst name=explain/
  str name=QParserExtendedDismaxQParser/str
  null name=altquerystring/
  null name=boostfuncs/
  arr name=filter_queries
  strprices:[5.00 TO 21.00]/str
  strforsaleinusa:true/str
  /arr
  arr name=parsed_filter_queries
  str
  ConstantScore(frange(currency(prices)):[500 TO 2100])
  /str
  strforsaleinusa:true/str
  /arr
  lst name=timing
  double name=time3.0/double
  lst name=prepare
  double name=time2.0/double
  lst name=org.apache.solr.handler.component.QueryComponent
  double name=time2.0/double
  /lst
  lst name=org.apache.solr.handler.component.FacetComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.MoreLikeThisComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.HighlightComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.StatsComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.SpellCheckComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.DebugComponent
  double name=time0.0/double
  /lst
  /lst
  lst name=process
  double name=time1.0/double
  lst name=org.apache.solr.handler.component.QueryComponent
  double name=time1.0/double
  /lst
  lst name=org.apache.solr.handler.component.FacetComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.MoreLikeThisComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.HighlightComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.StatsComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.SpellCheckComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.DebugComponent
  double name=time0.0/double
  /lst
  /lst
  /lst
  /lst
 
 
  If I run this same query with the filter, prices:[5.00 TO 99.00], then I
  get a result that includes the following field:
 
  arr name=prices
  str12.99,USD/str
  str14.99,USD/str
  str15.00,USD/str
  str25.00,USD/str
  /arr
 
 
  I can't figure out why this is not being returned with the first query.
  I'll try re-building the index with the prices field type set to float
  and see if that changes the behaviour.
 
  On Sat, Jun 30, 2012 at 6:49 PM, Erick Erickson erickerick...@gmail.com
 wrote:
 
  This works fine for me with 3.6, float fields and even on a currency
 type.
 
  I'm assuming a typo for 15.00.00 BTW.
 
  I admit I'm not all that familiar with the currency type, which I
 infer
  you're
  using given the USD bits. But I ran a quick test with currency types
 and
  it worked at least the way I ran it... But another 

deleteById commitWithin question

2012-07-05 Thread Jamie Johnson
I am running off of a snapshot taken 5/3/2012 of solr 4.0 and am
noticing some issues around deleteById when a commitWithin parameter
is included using SolrJ, specifically commit isn't executed.  If I
later just call commit on the solr instance I see the item is deleted
though.  Is anyone aware if this should work in that snapshot?


Re: deleteById commitWithin question

2012-07-05 Thread Yonik Seeley
On Thu, Jul 5, 2012 at 4:29 PM, Jamie Johnson jej2...@gmail.com wrote:
 I am running off of a snapshot taken 5/3/2012 of solr 4.0 and am
 noticing some issues around deleteById when a commitWithin parameter
 is included using SolrJ, specifically commit isn't executed.  If I
 later just call commit on the solr instance I see the item is deleted
 though.  Is anyone aware if this should work in that snapshot?


I thought I remembered something like this... but looking at the
commit log for DUH2, I don't see it.

/opt/code/lusolr4$ svn log
./solr/core/src/java/org/apache/solr/update/DirectUpdateHandler2.java
| less


r1357332 | yonik | 2012-07-04 12:23:09 -0400 (Wed, 04 Jul 2012) | 1 line

log DBQ reordering events

r1356858 | markrmiller | 2012-07-03 14:18:48 -0400 (Tue, 03 Jul 2012) | 1 line

SOLR-3587: After reloading a SolrCore, the original Analyzer is still
used rather than a new one

r1356845 | yonik | 2012-07-03 13:47:56 -0400 (Tue, 03 Jul 2012) | 1 line

SOLR-3559: DBQ reorder support

r1355088 | sarowe | 2012-06-28 13:51:38 -0400 (Thu, 28 Jun 2012) | 1 line

LUCENE-4172: clean up redundant throws clauses (merge from trunk)

r1348984 | hossman | 2012-06-11 15:46:14 -0400 (Mon, 11 Jun 2012) | 1 line

LUCENE-3949: fix license headers to not be javadoc style comments

r1343813 | rmuir | 2012-05-29 12:16:38 -0400 (Tue, 29 May 2012) | 1 line

create stable branch for 4.x releases

r1328890 | yonik | 2012-04-22 11:01:55 -0400 (Sun, 22 Apr 2012) | 1 line

SOLR-3392: fix search leak when openSearcher=false

r1328883 | yonik | 2012-04-22 09:58:00 -0400 (Sun, 22 Apr 2012) | 1 line

SOLR-3391: Make explicit commits cancel pending autocommits.


I'll try out trunk quick and see if it currently works.

-Yonik
http://lucidimagination.com


Re: deleteById commitWithin question

2012-07-05 Thread Jamie Johnson
Thanks for the fast response.  Yeah, it could be in either place
either SolrJ or on the Server side.

On Thu, Jul 5, 2012 at 4:47 PM, Yonik Seeley yo...@lucidimagination.com wrote:
 On Thu, Jul 5, 2012 at 4:29 PM, Jamie Johnson jej2...@gmail.com wrote:
 I am running off of a snapshot taken 5/3/2012 of solr 4.0 and am
 noticing some issues around deleteById when a commitWithin parameter
 is included using SolrJ

 Oh wait... I just realized you were talking about SolrJ specifically -
 so the issue may be there.

 -Yonik
 http://lucidimagination.com


RE: Update JSON format post 3.1?

2012-07-05 Thread Klostermeyer, Michael
Sorry...I found the answer in the comments of the previously mentioned Jira 
ticket.  Apparently the proposed solution differed from the final one (the 
doc structure key is not needed, apparently).

Mike


-Original Message-
From: Klostermeyer, Michael [mailto:mklosterme...@riskexchange.com] 
Sent: Thursday, July 05, 2012 12:55 PM
To: solr-user@lucene.apache.org
Subject: Update JSON format post 3.1?

Is there any official documentation on the JSON format Solr 3.5 expects when 
adding/updating documents via /update/json?  Most of the official documentation 
is 3.1, and I am of the understanding this changed in v3.2 
(https://issues.apache.org/jira/browse/SOLR-2496).  I believe I have the 
correct format, but I am getting an odd error:

The request sent by the client was syntactically incorrect (Expected: 
OBJECT_START but got ARRAY_START

My JSON format is as follows (simplified):
{add:
{doc:
[
{ID:987654321,Name:Steve Smith,ChildIDs:[3841]} ] } }

The idea is that I want to be able to send multiple documents within the same 
request, although in this example I am demonstrating only a single document. 
ChildIDs is defined as a multivalued field.

Thanks.

Mike Klostermeyer


Re: deleteById commitWithin question

2012-07-05 Thread Jamie Johnson
Ok, so some more context, hopefully this is useful.

I didn't think this was a SolrCloud issue but it appears to be.  I
have a simple 2 shard set up, I add 1 document which goes to shard 1.
I then issue a delete to shard 2.  The delete gets there and I see the
commit eventually being run in the log, but it doesn't appear to be
being distributed to shard 1.  If I manually call a server.commit()
the item is removed.

On Thu, Jul 5, 2012 at 4:56 PM, Jamie Johnson jej2...@gmail.com wrote:
 Thanks for the fast response.  Yeah, it could be in either place
 either SolrJ or on the Server side.

 On Thu, Jul 5, 2012 at 4:47 PM, Yonik Seeley yo...@lucidimagination.com 
 wrote:
 On Thu, Jul 5, 2012 at 4:29 PM, Jamie Johnson jej2...@gmail.com wrote:
 I am running off of a snapshot taken 5/3/2012 of solr 4.0 and am
 noticing some issues around deleteById when a commitWithin parameter
 is included using SolrJ

 Oh wait... I just realized you were talking about SolrJ specifically -
 so the issue may be there.

 -Yonik
 http://lucidimagination.com


Re: Solr 4.0 UI issue

2012-07-05 Thread Stefan Matheis
Hi, we are working on https://issues.apache.org/jira/browse/SOLR-3591 - so you 
may check the (raw) log output .. which may cause the UI showing this message.



On Thursday, July 5, 2012 at 7:19 PM, anarchos78 wrote:

 Greetings friends,
 I just discovered today that there is a new solr release (4.0 ALPHA). So, I
 give it a try. After setting it up (under Tomcat) I had the following error
 message:  
 “This interface requires that you activate the admin request handlers, add
 the following configuration to your solrconfig.xml:  
  
 requestHandler name=/admin/ class=solr.admin.AdminHandlers /”
  
 The above error popped up when I added the following in the solrconfig.xml:
  
 requestHandler name=/dataimport
 class=org.apache.solr.handler.dataimport.DataImportHandler
 lst name=defaults
 str name=configdata-config.xml/str
 /lst
 /requestHandler
  
 Does anybody know what it is wrong?
  
 Thank you in advance,
  
 Tom
 Greece
  
 *The solr folder structure:*
 +solr
 +conf
 +data
 +lib
 -contrib
 -dist
  
 *The solr.xml file:*
 ?xml version=1.0 encoding=UTF-8 ?
 solr persistent=false
 cores adminPath=/admin/cores defaultCoreName=collection1
 core name=collection1 instanceDir=. /
 /cores
 /solr
  
 *The solrconfig.xml file:*
 ?xml version=1.0 encoding=UTF-8 ?
 config
 luceneMatchVersionLUCENE_40/luceneMatchVersion
  
 lib dir=lib/dist/ regex=apache-solr-cell-\d.*\.jar /
 lib dir=lib/contrib/extraction/lib/ regex=.*\.jar /
  
 lib dir=lib/dist/ regex=apache-solr-clustering-\d.*\.jar /
 lib dir=lib/contrib/clustering/lib/ regex=.*\.jar /
  
 lib dir=lib/dist/ regex=apache-solr-dataimporthandler-\d.*\.jar /
  
  
 lib dir=lib/dist/ regex=apache-solr-langid-\d.*\.jar /
 lib dir=lib/contrib/langid/lib/ regex=.*\.jar /
  
 lib dir=lib/dist/ regex=apache-solr-velocity-\d.*\.jar /
 lib dir=lib/contrib/velocity/lib/ regex=.*\.jar /
  
 lib dir=lib/dist/
 regex=apache-solr-dataimporthandler-extras-\d.*\.jar /
  
 lib dir=lib/contrib/extraction/lib/ regex=tika-core-\d.*\.jar /
 lib dir=lib/contrib/extraction/lib/ regex=tika-parsers-\d.*\.jar /
 lib dir=/total/crap/dir/ignored /  
  
 dataDir${solr.data.dir:}/dataDir
  
 directoryFactory name=DirectoryFactory  
  
 class=${solr.directoryFactory:solr.StandardDirectoryFactory}/
  
 indexConfig
  
 /indexConfig
 jmx /
  
 updateHandler class=solr.DirectUpdateHandler2
  
 autoCommit  
 maxTime15000/maxTime  
 openSearcherfalse/openSearcher  
 /autoCommit
  
 updateLog
 str name=dir${solr.data.dir:}/str
 /updateLog
  
  
 /updateHandler
  
 query
  
 maxBooleanClauses1024/maxBooleanClauses
  
 filterCache class=solr.FastLRUCache
 size=512
 initialSize=512
 autowarmCount=0/
  
 queryResultCache class=solr.LRUCache
 size=512
 initialSize=512
 autowarmCount=0/
  
 documentCache class=solr.LRUCache
 size=512
 initialSize=512
 autowarmCount=0/
  
 enableLazyFieldLoadingtrue/enableLazyFieldLoading
  
 queryResultWindowSize20/queryResultWindowSize
  
 queryResultMaxDocsCached200/queryResultMaxDocsCached
  
 listener event=newSearcher class=solr.QuerySenderListener
 arr name=queries
  
 /arr
 /listener
 listener event=firstSearcher class=solr.QuerySenderListener
 arr name=queries
 lst
 str name=qstatic firstSearcher warming in solrconfig.xml/str
 /lst
 /arr
 /listener
  
 useColdSearcherfalse/useColdSearcher
  
 maxWarmingSearchers2/maxWarmingSearchers
  
 /query
  
 requestDispatcher handleSelect=false 
  
 requestParsers enableRemoteStreaming=true  
 multipartUploadLimitInKB=2048000 /
  
 httpCaching never304=true /
  
 /requestDispatcher
  
 requestHandler name=/dataimport
 class=org.apache.solr.handler.dataimport.DataImportHandler
 lst name=defaults
 str name=configdata-config.xml/str
 /lst
 /requestHandler
  
 requestHandler name=/select class=solr.SearchHandler
  
 lst name=defaults
 str name=echoParamsexplicit/str
 int name=rows10/int
 str name=dftext/str
 /lst
  
 /requestHandler
  
 requestHandler name=/query class=solr.SearchHandler
 lst name=defaults
 str name=echoParamsexplicit/str
 str name=wtjson/str
 str name=indenttrue/str
 str name=dftext/str
 /lst
 /requestHandler
  
 requestHandler name=/get class=solr.RealTimeGetHandler
 lst name=defaults
 str name=omitHeadertrue/str
 /lst
 /requestHandler
  
 requestHandler name=/browse class=solr.SearchHandler
 lst name=defaults
 str name=echoParamsexplicit/str
  
 str name=wtvelocity/str
 str name=v.templatebrowse/str
 str name=v.layoutlayout/str
 str name=titleSolritas/str
  
 str name=defTypeedismax/str
 str name=qf
 text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0 manu^1.1 cat^1.4
 /str
 str name=mm100%/str
 str name=q.alt*:*/str
 str name=rows10/str
 str name=fl*,score/str
  
 str name=mlt.qf
 text^0.5 features^1.0 name^1.2 sku^1.5 id^10.0 manu^1.1 cat^1.4
 /str
 str name=mlt.fltext,features,name,sku,id,manu,cat/str
 int name=mlt.count3/int
  
 str name=faceton/str
 str name=facet.fieldcat/str
 str name=facet.fieldmanu_exact/str
 str name=facet.queryipod/str
 str name=facet.queryGB/str
 str name=facet.mincount1/str
 str 

Re: Filtering a query by range returning unexpected results

2012-07-05 Thread Erick Erickson
For future reference, JIRA is here:
https://issues.apache.org/jira/browse/SOLR-3595

Erick

On Thu, Jul 5, 2012 at 4:19 PM, Andrew Meredith andymered...@gmail.com wrote:
 Thanks! That worked. I re-built the index with my prices field being of
 the tfloat type, and I am now able to perform range queries. I appreciate
 your help, Erick.

 On Tue, Jul 3, 2012 at 9:35 AM, Erick Erickson erickerick...@gmail.comwrote:

 OK, this appears to be something with the currency type. It works fine
 for
 regular float fields. I can't get the multiValued currency types to work
 with
 range queries. Don't quite know what I was doing when I thought they
 _did_ work.

 One work-around I think, if you are using a single currency USD might be
 to copy your price to a simple float field and do your range queries on
 that.

 I'm not at all sure that the currency type was ever intended to support
 multiValued=true. I don't know enough about the internals to know if
 it's even a good idea to try, but the current behavior could be improved
 upon.

 But it seems to me that one of two things should happen:
 1 the startup should barf if a currency type is multiValued (fail early)
 or
 2 currency should work when multiValued.

 Unfortunately, JIRA is down so I can't look to see if this is already a
 known
 issue or enter a JIRA if it isn't. I'll try to look later if it all
 comes back up.

 Best
 Erick

 On Mon, Jul 2, 2012 at 1:53 PM, Andrew Meredith andymered...@gmail.com
 wrote:
  Yep, that 15.00.00 was a typo.
 
  Here are the relevant portions of my schema.xml:
 
  types
  !-- CUT --
  fieldType name=currency class=solr.CurrencyField
 precisionStep=8
  defaultCurrency=USD currencyConfig=currency.xml /
  !-- CUT --
  /types
 
  fields
  !-- CUT --
  field name=prices type=currency indexed=true stored=true
  multiValued=true /
  !-- CUT --
  /fields
 
  And here is the output of a sample query with debugQuery=on appended:
 
  lst name=debug
  str name=rawquerystringFurtick/str
  str name=querystringFurtick/str
  str name=parsedquery
  +DisjunctionMaxQuery((subtitle:furtick | frontlist_flapcopy:furtick^0.5 |
  frontlist_ean:furtick^6.0 | author:furtick^3.0 | series:furtick^1.5 |
  title:furtick^2.0)) ()
  /str
  str name=parsedquery_toString
  +(subtitle:furtick | frontlist_flapcopy:furtick^0.5 |
  frontlist_ean:furtick^6.0 | author:furtick^3.0 | series:furtick^1.5 |
  title:furtick^2.0) ()
  /str
  lst name=explain/
  str name=QParserExtendedDismaxQParser/str
  null name=altquerystring/
  null name=boostfuncs/
  arr name=filter_queries
  strprices:[5.00 TO 21.00]/str
  strforsaleinusa:true/str
  /arr
  arr name=parsed_filter_queries
  str
  ConstantScore(frange(currency(prices)):[500 TO 2100])
  /str
  strforsaleinusa:true/str
  /arr
  lst name=timing
  double name=time3.0/double
  lst name=prepare
  double name=time2.0/double
  lst name=org.apache.solr.handler.component.QueryComponent
  double name=time2.0/double
  /lst
  lst name=org.apache.solr.handler.component.FacetComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.MoreLikeThisComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.HighlightComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.StatsComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.SpellCheckComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.DebugComponent
  double name=time0.0/double
  /lst
  /lst
  lst name=process
  double name=time1.0/double
  lst name=org.apache.solr.handler.component.QueryComponent
  double name=time1.0/double
  /lst
  lst name=org.apache.solr.handler.component.FacetComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.MoreLikeThisComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.HighlightComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.StatsComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.SpellCheckComponent
  double name=time0.0/double
  /lst
  lst name=org.apache.solr.handler.component.DebugComponent
  double name=time0.0/double
  /lst
  /lst
  /lst
  /lst
 
 
  If I run this same query with the filter, prices:[5.00 TO 99.00], then I
  get a result that includes the following field:
 
  arr name=prices
  str12.99,USD/str
  str14.99,USD/str
  str15.00,USD/str
  str25.00,USD/str
  /arr
 
 
  I can't figure out why this is not being returned with the first query.
  I'll try re-building the index with the prices field type set to float
  and see if that changes the behaviour.
 
  On Sat, Jun 30, 2012 at 6:49 PM, Erick Erickson erickerick...@gmail.com
 wrote:
 
  This works fine for me with 3.6, float fields and even on a currency
 type.
 
  I'm assuming a typo for 15.00.00 BTW.
 
  I admit I'm not all that familiar 

Re: deleteById commitWithin question

2012-07-05 Thread Jamie Johnson
Thanks Yonik, glad that it is not an issue on Trunk.  I'll see if our
group is interested in updating now to take advantage of this.  In the
mean time we just issue a commit after we do deletes and that seems to
be a sufficient workaround.

On Thu, Jul 5, 2012 at 6:05 PM, Yonik Seeley yo...@lucidimagination.com wrote:
 On Thu, Jul 5, 2012 at 5:12 PM, Jamie Johnson jej2...@gmail.com wrote:
 Ok, so some more context, hopefully this is useful.

 I didn't think this was a SolrCloud issue but it appears to be.  I
 have a simple 2 shard set up, I add 1 document which goes to shard 1.
 I then issue a delete to shard 2.  The delete gets there and I see the
 commit eventually being run in the log, but it doesn't appear to be
 being distributed to shard 1.  If I manually call a server.commit()
 the item is removed.

 OK, I just verified this for trunk / 4x.  It appears to work fine now,
 but I'm not exactly sure why it may not with the snapshot you have
 (much has changed  since).

 I followed the solr cloud wiki except that the path now needs to have
 collection1 in it, and I set the number of shards to 1 (with 2
 replicas).   I indexed the example docs, then queried both shards to
 verify the doc existed on both.  Then I deleted the doc through the
 non-leader via:
 curl http://localhost:7574/solr/update?commitWithin=6; -H
 'Content-type:application/json' -d '{delete: { id:SOLR1000}}'

 This was to test both the forwarding to the leader and the fowarding
 from the leader to the replica.
 After 60 seconds, the document was gone on both replicas.

 aside: If you use deleteByQuery, be aware that full distributed
 indexing support was just finished yesterday by SOLR-3559

 -Yonik
 http://lucidimagination.com


Re: highlighting and wildcards

2012-07-05 Thread Jamie Johnson
Looks like this issue has already been fixed, sorry for the static.  I
will update and try again.

On Thu, Jul 5, 2012 at 5:52 PM, Jamie Johnson jej2...@gmail.com wrote:
 I am executing a query on a multivalued text field that has the value
 of ZtestForJamie.  When I execute this query

 http://localhost:8501/solr/select?q=ztest*fl=subjecthl=true

 the hit that comes back is as follows

 emZtest/emForemZtestForJamie/em

 I am running a solr 4.0 snapshot from 5/3/2012.


Re: ExtendedDisMax Field Alias Question

2012-07-05 Thread Jamie Johnson
It's been some time since I've thought about this, but I wanted anyone
interested to know I created the following JIRA asking for this
feature.

https://issues.apache.org/jira/browse/SOLR-3598

On Sat, May 26, 2012 at 9:28 PM, Jamie Johnson jej2...@gmail.com wrote:
 Yeah cycles in general I agree are bad, but perhaps an option to also
 include the original field or special handling of the aliased field to
 support this.

 On Sat, May 26, 2012 at 3:26 PM, Jack Krupansky j...@basetechnology.com 
 wrote:
 That would create an alias loop, which is not supported.

 For example,

 http://localhost:8983/solr/select/?debugQuery=truedefType=edismaxf.person_first_name.qf=genre_sf.person_last_name.qf=idf.name.qf=name+person_first_name+person_last_nameq=name:smith

 in Solr 3.6 generates a 400 response status code with this exception:

 org.apache.lucene.queryParser.ParseException: Cannot parse 'name:smith ':
 Field aliases lead to a cycle

 Maybe what you would like is an enhancement to permit an explicit
 refererence to the underlying field rather than the alias in an alias
 definition, like:

 f.name.qf=field.name+person_first_name+person_last_name

 -- Jack Krupansky

 -Original Message- From: Jamie Johnson
 Sent: Friday, May 25, 2012 8:37 PM
 To: solr-user@lucene.apache.org
 Subject: ExtendedDisMax Field Alias Question


 I was wondering if someone could explain if the following is supported
 with the current EDisMax Field Aliasing.

 I have a field like person_name which exists in solr, we also have 2
 other fields named person_first_name and person_last_name.  I would
 like to allow queries for person_name to be aliased as person_name,
 person_first_name and person_last_name.  Is this allowed or does the
 alias need to not appear in the list of fields to be aliased to (I
 remember seeing something about aliases to other aliases is allowed)?
 I could obviously create a purely virtual field which aliases all 3
 but it would be nice if the parser could support this case.


Re: Solr facet multiple constraint

2012-07-05 Thread davidbougearel
Well thanks for your answer, in fact i've written what the QueryResponse
return as the solr query here is my real solr query before use the
executeQuery :

q=service%3A1+AND+publicationstatus%3ALIVEsort=publishingdate+descfq=%7B%21ex%3Ddt%7D%28%28%28user%3A10%29%29%29facet.field=%7B%21tag%3Ddt%7Duserfacet=truefacet.mincount=1

which is the same as my first post without the 'wt=javabin' and  instead of
commas.

Could you please see if there is something wrong for you ?

Best regards, 

David.

--
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-facet-multiple-constraint-tp3992974p3993408.html
Sent from the Solr - User mailing list archive at Nabble.com.


Regression of JIRA 1826?

2012-07-05 Thread Jamie Johnson
I just upgraded to trunk to try to fix an issue I was having with the
highlighter described in JIRA 1826, but it appears that this issue
still exists on trunk.  I'm running the following query

subject:ztest*

subject is a text field (not multivalued) and the return in highlighting is

emZTest/emForemZTestForJamie/em

the actual stored value is ZTestForJamie.  Is anyone else experiencing this?


Re: Dismax Question

2012-07-05 Thread Steve Fatula
It turns out that Solr 3.5.0 does not have the dismax issue, so, we have 
reverted. Hopefully, the bug will be fixed.