how to range search on the field which contains multiple decimal point (eg: 2.5.0.4)

2016-05-04 Thread Santhosh Sheshasayanan
Hi, I have issue in my server. As I stated in the subject I want to do range search query on the field (eg: filed name is "version") which contains value like (2.5.0.1, 2.5.0.4 and 2.5.0.10 etc). When I do range search on the "version" field with criteria [* TO 2.5.0.5], it gave me all the

Advice to add additional non-related fields to a collection or create a subset of it?

2016-05-04 Thread Derek Poh
Hi We have a "product" collection and a "supplier" collection. The "product" collection contains products information and "supplier" collection contains the product's suppliers information. We have a subsidiary page that query on "product" collection for the search. The display result include

RE: Query String Limit

2016-05-04 Thread Prasanna S. Dhakephalkar
Hi We had increased the maxBooleanClauses to a large number, but it did not work Here is the query http://localhost:8983/solr/collection1/select?fq=record_id%3A(604929+504197+ 500759+510957+624719+524081+544530+375687+494822+468221+553049+441998+495212

Re: Migrating from Solr 5.4 to Solr 6.0

2016-05-04 Thread Zheng Lin Edwin Yeo
Thank you. That would save us quite alot of time, as we are worried that our current index will not be compatible with the new BM25 scoring algorithm. Regards, Edwin On 4 May 2016 at 19:50, Markus Jelsma wrote: > No, you don't need to reindex. > M. > >

Using MoreLikeThis for multiple documents/keywords

2016-05-04 Thread Zheng Lin Edwin Yeo
Hi, I would like to find out, if it is possible to use MoreLikeThis to get the response and interesting terms based on 2 or more multiple documents or keywords, by adding the AND or OR parameters in the query like what we do during search? For example:

Re: MoreLikeThis Component - how to get fields of documents

2016-05-04 Thread Zheng Lin Edwin Yeo
Hi Jan, Which version of Solr are you using? Regards, Edwin On 26 April 2016 at 23:46, Dr. Jan Frederik Maas < jan.m...@sub.uni-hamburg.de> wrote: > Hello, > > I want to use the moreLikeThis Component to get similar documents from a > sharded SOLR. This works quite well except for the fact

Re: Solr cloud 6.0.0 with ZooKeeper 3.4.8 Errors

2016-05-04 Thread Susheel Kumar
Thanks, Nick & Hoss. I am using the exact same machine, have wiped out solr 5.5.0 and installed solr-6.0.0 with external ZK 3.4.8. I checked the File Description limit for user solr, which is 12000 and increased to 52000. Don't see "too many files open..." error now in Solr log but still Solr

Re: Bug in Solr 6 dynamic-fields?

2016-05-04 Thread Alexandre Rafalovitch
I've just answered this on SO, but I think the reason is quite confusing and hope others can comment here Basically, the issue is that "string" field type has docValues enabled and therefore with schema 1.6, any field inherited from it will be effectively both searchable and returnable, even if

RE: Integrating grobid with Tika in solr

2016-05-04 Thread Allison, Timothy B.
Y, integrating Tika is non-trivial. I think Uwe adds the dependencies with great care by hand by carefully looking at the dependency tree in Maven and making sure there weren't any conflicts. -Original Message- From: Shawn Heisey [mailto:apa...@elyograg.org] Sent: Wednesday, May 4,

Re: Solr cloud 6.0.0 with ZooKeeper 3.4.8 Errors

2016-05-04 Thread Chris Hostetter
: Thanks, Nick. Do we know any suggested # for file descriptor limit with : Solr6? Also wondering why i haven't seen this problem before with Solr 5.x? are you running Solr6 on the exact same host OS that you were running Solr5 on? even if you are using the "same OS version" on a diff

getZkStateReader() returning NULL

2016-05-04 Thread Boman
I am attempting to check for existence of a collection prior to creating a new one with that name, using Solrj: System.out.println("Checking for existence of collection..."); ZkStateReader zkStateReader = this.server.getZkStateReader(); zkStateReader.updateClusterState();

Re: OOM script executed

2016-05-04 Thread Chris Hostetter
: You could, but before that I'd try to see what's using your memory and see : if you can decrease that. Maybe identify why you are running OOM now and : not with your previous Solr version (assuming you weren't, and that you are : running with the same JVM settings). A bigger heap usually means

Re: Solr cloud 6.0.0 with ZooKeeper 3.4.8 Errors

2016-05-04 Thread Nick Vasilyev
Not sure about your environment so it's hard to say why you haven't ran into this issue before. As for the suggested limit, I am not sure, it would depend on your system and if you really want to limit it. I personally just jack it up to 5. On Wed, May 4, 2016 at 6:13 PM, Susheel Kumar

Re: Solr cloud 6.0.0 with ZooKeeper 3.4.8 Errors

2016-05-04 Thread Susheel Kumar
Thanks, Nick. Do we know any suggested # for file descriptor limit with Solr6? Also wondering why i haven't seen this problem before with Solr 5.x? On Wed, May 4, 2016 at 4:54 PM, Nick Vasilyev wrote: > It looks like you have too many open files, try increasing the

Re: Solr cloud 6.0.0 with ZooKeeper 3.4.8 Errors

2016-05-04 Thread Nick Vasilyev
It looks like you have too many open files, try increasing the file descriptor limit. On Wed, May 4, 2016 at 3:48 PM, Susheel Kumar wrote: > Hello, > > I am trying to setup 2 node Solr cloud 6 cluster with ZK 3.4.8 and used the > install service to setup solr. > > After

Solr cloud 6.0.0 with ZooKeeper 3.4.8 Errors

2016-05-04 Thread Susheel Kumar
Hello, I am trying to setup 2 node Solr cloud 6 cluster with ZK 3.4.8 and used the install service to setup solr. After launching Solr Admin Panel on server1, it looses connections in few seconds and then comes back and other node server2 is marked as Down in cloud graph. After few seconds its

Faceting and Grouping Performance Degradation in Solr 5

2016-05-04 Thread Solr User
I recently was attempting to upgrade from Solr 4.8.1 to Solr 5.4.1 but had to abort due to average response times degraded from a baseline volume performance test. The affected queries involved faceting (both enum method and default) and grouping. There is a critical bug

Re: Integrating grobid with Tika in solr

2016-05-04 Thread Shawn Heisey
On 5/4/2016 9:21 AM, Betsey Benagh wrote: > I’m feeling particularly dense, because I don’t see any Tika jars in > WEB-INF/lib: Oops. Sorry about that, I forgot that it's all contrib. That's my mistake, not yours. The Tika jars are in contrib/extraction/lib, along with a very large number of

Bug in Solr 6 dynamic-fields?

2016-05-04 Thread Tech Id
Hi, We are unable to resolve a problem with dynamic fields in Solr 6. The question and details can be found on stack-overflow at http://stackoverflow.com/questions/37014345/unable-to-add-new-dynamic-fields-in-solr-6-0/37018450#37018450 If its a real bug, then we can file a JIRA for the same.

Re: Integrating grobid with Tika in solr

2016-05-04 Thread Betsey Benagh
As a workaround, I’m trying to run Grobid on my files, and then import the corresponding XML into Solr. I don’t see any errors on the post: bba0124$ bin/post -c lrdtest ~/software/grobid/out/021002_1.tei.xml /Library/Java/JavaVirtualMachines/jdk1.8.0_71.jdk/Contents/Home/bin/java -classpath

Re: What does the "Max Doc" means in Admin interface?

2016-05-04 Thread John Bickerstaff
Max doc is the total amount of documents in the collection INCLUDING the ones that have been deleted but not actually removed. Don't worry, deleted docs are not used in search results. Yes, you can change the number by "optimizing" (see the button) but this does take time and bandwidth so use it

ReversedWildcardFilterFactory question

2016-05-04 Thread Susheel Kumar
Hello, I wanted to confirm that using below type for fields where user *may also* search for leading wildcard, is a good solution and edismax query parser would automatically reverse the query string in case of leading wildcard search e.g. q:"text:*plane" would automatically be reversed by

Re: SOLR edismax and mm request parameter

2016-05-04 Thread ND
If I am understanding you correctly, it sounds like you are looking for an phrase match with a really large query slop parameter (qs, https://cwiki.apache.org/confluence/display/solr/The+DisMax+Query+Parser#TheDisMaxQueryParser-Theqs%28QueryPhraseSlop%29Parameter), I believe the old way of doing

Re: Integrating grobid with Tika in solr

2016-05-04 Thread Betsey Benagh
I’m feeling particularly dense, because I don’t see any Tika jars in WEB-INF/lib: antlr4-runtime-4.5.1-1.jar asm-5.0.4.jar asm-commons-5.0.4.jar commons-cli-1.2.jar commons-codec-1.10.jar commons-collections-3.2.2.jar commons-configuration-1.6.jar commons-exec-1.3.jar

Re: Integrating grobid with Tika in solr

2016-05-04 Thread Shawn Heisey
On 5/4/2016 8:38 AM, Betsey Benagh wrote: > Thanks, I¹m currently using 5.5, and will try upgrading to 6.0. > > > On 5/4/16, 10:37 AM, "Allison, Timothy B." wrote: >> Y. Solr 6.0.0 is shipping with Tika 1.7. Grobid came in with Tika 1.11. Just upgrading to 6.0.0 isn't

Re: Integrating grobid with Tika in solr

2016-05-04 Thread Betsey Benagh
Thanks, I¹m currently using 5.5, and will try upgrading to 6.0. On 5/4/16, 10:37 AM, "Allison, Timothy B." wrote: >Y. Solr 6.0.0 is shipping with Tika 1.7. Grobid came in with Tika 1.11. > >-Original Message- >From: Allison, Timothy B. [mailto:talli...@mitre.org]

RE: Integrating grobid with Tika in solr

2016-05-04 Thread Allison, Timothy B.
Y. Solr 6.0.0 is shipping with Tika 1.7. Grobid came in with Tika 1.11. -Original Message- From: Allison, Timothy B. [mailto:talli...@mitre.org] Sent: Wednesday, May 4, 2016 10:29 AM To: solr-user@lucene.apache.org Subject: RE: Integrating grobid with Tika in solr I think Solr is using

RE: Integrating grobid with Tika in solr

2016-05-04 Thread Allison, Timothy B.
I think Solr is using a version of Tika that predates that addition of the Grobid parser. You'll have to add that manually somehow until Solr upgrades to Tika 1.13 (soon to be released...I think). SOLR-8981. -Original Message- From: Betsey Benagh [mailto:betsey.ben...@stresearch.com]

Re: OOM script executed

2016-05-04 Thread Shawn Heisey
On 5/3/2016 11:58 PM, Bastien Latard - MDPI AG wrote: > Thank you for your email. > You said "have big caches or request big pages (e.g. 100k docs)"... > Does a fq cache all the potential results, or only the ones the query > returns? > e.g.: select?q=*:*=bPublic:true=10 > > => with this query, if

Re: solr.ICUCollationField class on cloudera search solr

2016-05-04 Thread Shawn Heisey
On 5/4/2016 3:53 AM, tkg_cangkul wrote: > i have check the library >> >> /opt/cloudera/parcels/CDH/lib/solr/solr-analysis-extras-4.10.3-cdh5.7.0.jar >>and there is solr.ICUCollationField class. but why i still have this >>error message? >> pls help You may be able to get this

Re: Integrating grobid with Tika in solr

2016-05-04 Thread Betsey Benagh
Grobid runs as a service, and I’m (theoretically) configuring Tika to call it. >From the Grobid wiki, here are instructions for integrating with Tika >application: First we need to create the GrobidExtractor.properties file that points to the Grobid REST Service. My file looks like the

Re: Integrating grobid with Tika in solr

2016-05-04 Thread Shawn Heisey
On 5/4/2016 7:15 AM, Betsey Benagh wrote: > (X-posted from stack overflow) > > This feels like a basic, dumb question, but my reading of the documentation > has not led me to an answer. > > > i'm using Solr to index journal articles. Using the out-of-the-box > configuration, it indexed the

Nodes appear twice in state.json

2016-05-04 Thread Markus Jelsma
Hi - we've just upgraded a development environment from 5.5 to Solr 6.0. After the upgrade, which went fine, we see two replica's appear twice in the cloud view (see below), both being leader. We've seen this happen before on some older 5.x versions. Is there a Jira issue i am missing? An

Integrating grobid with Tika in solr

2016-05-04 Thread Betsey Benagh
(X-posted from stack overflow) This feels like a basic, dumb question, but my reading of the documentation has not led me to an answer. i'm using Solr to index journal articles. Using the out-of-the-box configuration, it indexed the text of the documents, but I'm looking to use Grobid to

Re: Solr 6 / Solrj RuntimeException: First tuple is not a metadata tuple

2016-05-04 Thread Kevin Risden
> > java.sql.SQLException: java.lang.RuntimeException: First tuple is not a > metadata tuple > That is a client side error message meaning that the statement couldn't be handled. There should be better error handling around this, but its not in place currently. And on Solr side, the logs seem

RE: Migrating from Solr 5.4 to Solr 6.0

2016-05-04 Thread Markus Jelsma
No, you don't need to reindex. M. -Original message- > From:Zheng Lin Edwin Yeo > Sent: Wednesday 4th May 2016 13:27 > To: solr-user@lucene.apache.org > Subject: Migrating from Solr 5.4 to Solr 6.0 > > Hi, > > Would like to find out, do we need to re-index our

MoreLikeThis (MLT) search

2016-05-04 Thread Zheng Lin Edwin Yeo
Hi, Would like to find out, must the fieldType be indexing with string, before we can perform a MoreLikeThis (MLT) query? Currently, my indexes are indexed with the HMMChineseTokenizer, so will it work well for MLT query? Below is my configuration for the fileType which I'm planning to execute

Migrating from Solr 5.4 to Solr 6.0

2016-05-04 Thread Zheng Lin Edwin Yeo
Hi, Would like to find out, do we need to re-index our document when we migrate from Solr 5.4 to Solr 6.0 because of the change in scoring algorithm to BM25? Regards, Edwin

Re: SOLR edismax and mm request parameter

2016-05-04 Thread Mark Robinson
Thanks for the mail Jaques. I have a doubt here. When we use q.op=AND what I understood is, ALL query terms should be present any where across the various "qf" fields ie all of the query terms need not be present in one single field, but just need to be present for sure among the various qf

Re: Results of facet differs with change in facet.limit.

2016-05-04 Thread Modassar Ather
The "val1" is same for both the test with limit 100 and 200 so the following is true. limit=100 1225 1082 1076 limit=200 1366 1321 1315 This I have noticed irrespective of facet.limit too. Please refer to my previous mail for the example. Thanks, Modassar On Wed, May 4, 2016 at 3:01 PM, Toke

Re: SOLR edismax and mm request parameter

2016-05-04 Thread Jacques du Rand
Sorry I meant "Ahmet Arslan" answer :) On 4 May 2016 at 11:56, Jacques du Rand wrote: > Although Mark Robinson's answer is correct you are now using the DISMAX > not the Edismax parser... > You can also play around with changing q.op parameter to 'AND' > > > > On 4

Re: SOLR edismax and mm request parameter

2016-05-04 Thread Jacques du Rand
Although Mark Robinson's answer is correct you are now using the DISMAX not the Edismax parser... You can also play around with changing q.op parameter to 'AND' On 4 May 2016 at 11:40, Mark Robinson wrote: > Thanks much Ahmet! > > I will try that out. > > Best, >

Re: solr.ICUCollationField class on cloudera search solr

2016-05-04 Thread tkg_cangkul
hi Ahmet thx for your reply. i've try using full qualified class name as your suggestion but it still failed. error On 04/05/16 16:34, Ahmet Arslan wrote: Hi, Sometimes using full qualified class name works: using org.apache.x.y..z.ICUCollationField instead of solr.ICUCollationField Ahmet

Re: SOLR edismax and mm request parameter

2016-05-04 Thread Mark Robinson
Thanks much Ahmet! I will try that out. Best, Mark On Tue, May 3, 2016 at 11:53 PM, Ahmet Arslan wrote: > Hi Mark, > > You could do something like this: > > _query_:{!dismax qf='field1' mm='100%' v=$qq} > OR > _query_:{!dismax qf='field2' mm='100%' v=$qq} > OR >

Re: solr.ICUCollationField class on cloudera search solr

2016-05-04 Thread Ahmet Arslan
Hi, Sometimes using full qualified class name works: using org.apache.x.y..z.ICUCollationField instead of solr.ICUCollationField Ahmet On Wednesday, May 4, 2016 11:13 AM, tkg_cangkul wrote: hi i'm using solr in cloudera. when i try create core i've got this error

Re: Results of facet differs with change in facet.limit.

2016-05-04 Thread Toke Eskildsen
On Mon, 2016-05-02 at 15:53 +0530, Modassar Ather wrote: > E.g. > Query : text_field:term=f=100 > Result : > 1225 > 1082 > 1076 > > Query : text_field:term=f=200 > 1366 > 1321 > 1315 Is the "val1" in your limit=100 test the same term as your "val1" in your limit=200-test? Or to phrase it

Re: Include and exclude feature with multi valued fileds

2016-05-04 Thread Anil
Hi Ahmet, in my example DOC 3 also has id 2 (typo mistake). i am using edismax query parser. i will try the query you suggested. Regard, Anil On 4 May 2016 at 12:28, Ahmet Arslan wrote: > > > Hi Anil, > > It is weird that your query retrieves docID=2, it has not

Solr 6 / Solrj RuntimeException: First tuple is not a metadata tuple

2016-05-04 Thread deniz
I am trying to go through the steps here to start playing with the new api, but I am getting: java.sql.SQLException: java.lang.RuntimeException: First tuple is not a metadata tuple at

Re: Include and exclude feature with multi valued fileds

2016-05-04 Thread Ahmet Arslan
Hi Anil, It is weird that your query retrieves docID=2, it has not Facebook at all. What query parser are you using? Please try unary operators and without using quotes. q={!lucene} +customers:facebook -customers:google If I am not wrong above query should do the trick. But I didn't

Re: Results of facet differs with change in facet.limit.

2016-05-04 Thread Modassar Ather
Thanks Erick for your response. I checked with distrib=false. I tried with a smaller result set. *Search* E.g. text_field:term AND f:val1 Number of matches : 49 *Facet:* (distrib=true) text_field:term AND f:val1 *Result* Shard1 : 47 *Facet: *(distrib=false) text_field:term AND f:val1=false